Skip to content
Diosh Lequiron
AI & Technology13 min read

AI Ethics for Practitioners: Beyond the Principles

AI ethics principles describe what systems should be. Practitioners need operational answers to five specific questions that determine whether an AI system is actually accountable, transparent, and fair.

The standard AI ethics curriculum teaches four principles: fairness, accountability, transparency, and explainability. These principles are real and consequential. They correctly identify the properties that responsible AI systems should have. They are also, as operational guidance, nearly useless.

The gap is this: principles describe what a system should be, not what a practitioner should do when making specific decisions under specific constraints. A practitioner responsible for an AI deployment does not need to be told that AI systems should be fair. They need to know how to evaluate whether the system they are about to deploy is fair enough to deploy, what to do when the model performs well on aggregate metrics but produces disparate outcomes for a specific subgroup, and who is accountable when an AI-generated recommendation harms someone who did not consent to being included in the system''s decision-making scope.

These are operational questions. Principles do not answer them. Practitioners who have internalized the ethics principles but have not been given operational guidance make decisions that are locally reasonable and systematically harmful — not because they lack ethical awareness, but because the gap between principle and decision has not been bridged.

What follows is an attempt to bridge that gap: five practitioner-level questions that operationalize AI ethics, a framework for embedding ethical requirements into system design rather than adding them as review steps, and a description of the failure modes that occur when ethics is treated as a compliance activity rather than a design constraint.


Why Principles Fail at the Point of Decision

Consider the fairness principle in a concrete case. An organization is deploying an AI model to prioritize service requests. The model performs well on standard accuracy metrics across the full dataset. A practitioner reviewing the deployment asks: is this fair?

The fairness principle says: the model should treat similarly situated individuals similarly. But this does not tell the practitioner whether the model meets that standard. To answer the actual question, the practitioner needs to know: what groups are relevant to assess? What is the right fairness metric — equal accuracy rates across groups, equal false positive rates, equal false negative rates, or some other criterion? These fairness metrics are mathematically incompatible in many real settings; optimizing for one worsens another. Who decides which metric applies?

None of these decisions can be derived from the fairness principle itself. They require domain knowledge, organizational context, a defined population of affected parties, and an explicit choice about which harms are most important to avoid. The practitioner who is told to be fair and asked to make these decisions without further guidance will make them based on whatever is easiest to measure, which is almost never what fairness actually requires.

The same pattern holds for the other principles. Accountability tells you that someone should be responsible for AI decisions. It does not tell you who, at what granularity of decision, under what escalation conditions, or what "responsible" means when an AI system makes thousands of decisions per day that no human reviews individually. Transparency tells you AI use should be disclosed. It does not tell you to whom, in what format, at what point in the interaction, or what disclosure looks like when the AI is a backend component of a service the user experiences as human-mediated.

The principles are correct. They are not sufficient. Operational guidance requires something more specific.


Five Questions That Operationalize AI Ethics

These five questions are not a complete ethics framework. They are the minimum set of questions that every AI deployment decision should be able to answer before the system goes into production. Inability to answer them is a deployment blocker.

Question 1: Who is accountable for decisions this AI system produces?

Accountability in AI systems requires more specificity than "the organization is accountable" or "the AI team is accountable." Accountability needs to attach to specific decisions at an appropriate level of resolution, and it needs to attach to a person or a role, not a team or a process.

The accountability question has three components. First: who is accountable for the aggregate performance of the system — its error rates, its fairness properties, its overall reliability? This is a design and oversight accountability that typically belongs to whoever owns the AI system. Second: who is accountable for any specific harmful decision the system produces? This requires identifying, in advance, who reviews complaints, who investigates identified harms, and who has authority to override or invalidate AI-generated decisions when warranted. Third: what is the escalation path when the accountable party for specific decisions cannot resolve a situation — when a complaint reveals a systematic problem rather than an isolated error?

These three accountability questions have different answers, and they should all be answered before deployment. "We will figure it out" is not an answer. It is a commitment to accountability theater.

Question 2: How will affected parties know AI was involved in a decision about them?

The transparency principle requires disclosure. Operationally, disclosure design requires answering: who needs to know (all people affected by any AI output, or only those where AI played a determinative role), when they need to know (before the decision is made, with the decision communication, on request, always), what they need to know (that AI was involved, what the AI did, what data was used, how they can challenge the decision), and in what format (plain language accessible to a non-technical audience, specific enough to be meaningful).

The disclosure design also needs to account for cases where the organization''s business model creates incentives against disclosure — where telling users that a recommendation is AI-generated reduces their engagement or willingness to pay. This is a real tension, and the ethical requirement does not disappear because disclosure is commercially inconvenient. The disclosure design needs to resolve this tension explicitly, not defer it.

For practitioners implementing disclosure: the test is whether a reasonable person who was affected by an AI-generated decision would consider themselves adequately informed about AI''s role if they were given the disclosure as designed. Not whether legal counsel considers the disclosure sufficient for liability purposes. Not whether the disclosure is technically accurate. Whether a person in the position of someone affected by the decision would consider themselves adequately informed.

Question 3: What is the process for affected parties to challenge AI-generated decisions?

Accountability and transparency create the architecture for challenge processes. The challenge process itself requires independent specification.

A challenge process needs: a defined entry point (where does someone go to challenge a decision), a defined scope (what kinds of decisions can be challenged, and what does a successful challenge produce — reconsideration, reversal, compensation, or only an explanation), a defined timeline (how long does the process take, what are the interim protections for the challenger while the process is underway), and a defined escalation path (if the organization''s internal challenge process does not resolve the challenge to the challenger''s satisfaction, what external mechanisms exist).

In regulated contexts, challenge processes are often legally required, and the legal requirements provide some of the architecture. In unregulated contexts, the organization must design the challenge process explicitly or effectively deny affected parties any remedy. Absence of a challenge process is an ethical design choice — it is not a neutral omission.

Question 4: Who monitors for disparate impact, and at what cadence?

Fairness at deployment is not a guarantee of fairness over time. Distribution shift — the gradual divergence between the conditions the model was trained on and the conditions it is deployed in — can introduce disparate impact that was not present at launch. User behavior adaptation, population composition changes, and changes in the underlying domain all affect model behavior over time.

Monitoring for disparate impact requires: defined subgroups to monitor (which means the organization has committed to monitoring fairness along specific dimensions — it cannot monitor all possible dimensions simultaneously, so it must choose), defined metrics for each subgroup (and a defined threshold at which observed disparity triggers review), a defined owner for the monitoring function (someone who is responsible for running the analyses, interpreting results, and escalating when thresholds are exceeded), and a defined cadence (how frequently monitoring occurs — which should be driven by the expected rate of distribution shift in the domain, not by operational convenience).

Organizations that do not specify monitoring design before deployment are making a commitment to monitor when something goes wrong — which is when monitoring is too late to prevent harm and most useful for investigating a crisis that has already occurred.

Question 5: What are the boundaries beyond which human review is mandatory?

Every AI system has domains where its reliability is high enough that automation without case-by-case human review is appropriate, and domains where its reliability is insufficient, the stakes are high enough, or the accountability requirements are strong enough that human review is mandatory regardless of efficiency cost.

These boundaries need to be defined explicitly as design parameters, not discovered empirically after deployment. Defining them requires: identifying the decision types or conditions under which human review is triggered (high-value decisions, decisions affecting vulnerable populations, decisions where the model confidence is below a threshold, decisions that affect access to essential services), defining what human review means (nominal review by someone who has received the AI recommendation is not the same as independent review by someone who has not), and defining what "human review" produces (a record of who reviewed, what they considered, and what they decided — not just an approval flag).

These five questions are not exhaustive. They are the questions whose answers are most likely to be absent in AI deployments that produce ethical failures. An organization that can answer all five specifically and operationally has done meaningful ethics work. An organization that answers them with "we follow best practices" or "we comply with applicable law" has not.


Embedding Ethical Requirements in System Design

The standard model for AI ethics in organizations is a review process: a system is designed, built, and prepared for deployment, and then an ethics review is conducted to identify problems before launch. This model has a structural flaw. It positions ethics as an external check on a system that has already been designed — and many of the most consequential ethical choices are made in system design, not in pre-launch review.

The choice of training data is an ethical choice. The selection of the optimization target is an ethical choice. The design of the feedback mechanism is an ethical choice. The choice of which subgroups to evaluate fairness against is an ethical choice. The decision about when to involve humans in the decision loop is an ethical choice. All of these choices are made during system design. A pre-launch ethics review that evaluates the system as designed cannot change these choices without rebuilding significant portions of the system, which is rarely done.

The alternative is to embed ethical requirements as design constraints that operate throughout the system development process, not as a review that happens after design is complete.

In practice, this means: defining accountability, disclosure, challenge, monitoring, and human review requirements before system architecture is specified; treating these requirements as constraints that shape technical design rather than as properties to be verified after design; and including the owners of these ethical requirements — the people who will operate the accountability process, run the monitoring, and respond to challenges — as stakeholders in design decisions, not as reviewers of completed design.

The difference between ethics as design constraint and ethics as compliance review is the difference between building a system that is structured to support accountability and building a system that is checked for accountability problems after the fact. Systems built with accountability as a design constraint look different from systems built without it — they have different data schemas, different logging structures, different decision boundaries, different escalation paths. These differences cannot be added cheaply after the fact.


The Failure Modes of Ethics as Compliance Activity

When AI ethics is treated primarily as a compliance activity — something done to satisfy regulators, auditors, or public expectations — characteristic failure modes emerge. These are not edge cases. They appear consistently in organizations whose ethics processes are compliance-oriented rather than design-oriented.

The documentation failure: The organization has documented its AI ethics principles and can demonstrate that it has conducted ethics reviews. What it cannot demonstrate is that the reviews changed anything about system design, or that the ethical requirements identified in review were implemented in the deployed system. The documentation exists as evidence of process, not as evidence of outcome.

The metrics displacement failure: The organization monitors AI system performance using defined metrics. It reports compliance with those metrics. The metrics are not measuring what the ethical requirements actually require — they are measuring what is easy to measure. A fairness metric that shows equal accuracy rates by demographic group may be compatible with severely disparate false positive rates that produce different types of harm for different groups. The organization is compliant with its metrics. It is not meeting its ethical obligations.

The accountability diffusion failure: Multiple parties are listed as having accountability for AI system performance. No specific person or role is accountable for any specific outcome. When harm occurs, accountability is contested between parties, each of whom is technically correct that they were not solely responsible. The harm is documented. Accountability is never established. The failure mode that produced the harm is not addressed.

The challenge process fiction: The organization has a defined process for challenging AI-generated decisions. In practice, the process has a response time measured in months, produces outcomes that rarely result in changed decisions, and is not accessible to the affected populations most likely to need it. The existence of the challenge process is used to demonstrate that affected parties have recourse. The actual function of the challenge process is to absorb complaints without producing accountability.

The ethics review timing failure: Ethics reviews are conducted before deployment and before major changes. The ethical characteristics of a system change over time as its deployment context changes — new user populations, new use cases, distribution shift, regulatory change. The organization''s ethics review process is not calibrated to detect drift. The ethics review from the original deployment is treated as continuing validation of a system that has materially changed since it was reviewed.

These failure modes are not the result of bad intentions. They are the predictable result of a compliance orientation that treats ethics as process to be documented rather than outcomes to be achieved. The shift from compliance orientation to design orientation is not a philosophical shift — it is a change in how the organization structures its AI development process, its accountability architecture, and its monitoring systems.


The Practitioner''s Position

Practitioners implementing AI systems in organizations typically do not control the strategic decisions about which systems to build or which use cases to pursue. They operate within constraints set by those decisions. What practitioners control is the implementation — the specific choices about system design, data handling, decision boundaries, disclosure formats, and monitoring cadence that shape how an AI system actually functions in deployment.

These implementation choices are where AI ethics operates in practice. A practitioner who treats the five accountability questions as real requirements — who refuses to deploy a system until accountability is specified, who pushes back on disclosure designs that are technically accurate but meaningfully opaque, who insists on defined monitoring before launch rather than monitoring promised after launch — is doing AI ethics work in the place where it actually affects outcomes.

This is not comfortable. It creates conflict with schedules, with cost structures, and with stakeholders whose primary interest is deployment, not governance. The discomfort is precisely what distinguishes genuine AI ethics practice from compliance theater. The principles describe the outcome. The practitioner''s operational questions, rigorously applied, are what produce it.

ShareTwitter / XLinkedIn

Explore more

← All Writing