Skip to content

Latest commit

 

History

History
145 lines (99 loc) · 10.1 KB

Deontological Ethics By Monotonicity Shape Constraints.md

File metadata and controls

145 lines (99 loc) · 10.1 KB

Paper

  • Title: Deontological Ethics By Monotonicity Shape Constraints
  • Authors: Serena Wang, Maya Gupta (Google Research)
  • Link: http://proceedings.mlr.press/v108/wang20e/wang20e.pdf
  • Platform: International Conference on Artificial Intelligence and Statistics (AISTATS)
  • Year: 2020

Summary

What?

  • Non-linear ML models can violate social norms or deontological ethical principles.
  • Incorporating monotone behavior in ML models via shape constraints can help in overcoming this.
  • Moreover, these deontological constraints can be related to commonly known notions of statistical fairness (parity and equal opportunity).
  • In general, monotonicity shape constraints helps accelerate progress towards a trustworthy and responsible AI.

Monotonicity Fairness Constraints

  • Unfair Penalization (UP): There may be inputs that a responsible model may reward but should never penalize.
  • Favor the less fortunate (FLF): There may be inputs that help us identify the less fortunate and favor them if there are no other relevant differences.

How?

  • UP and FLF are tackled by constraining the ML model to only respond positively to relevant inputs if all other inputs are fixed. Such monotonicity shape restrictions are referred to ceterus paribus in nature.
  • Focus is on ML models that produce a score which is used to determine some benefit (regression / classification), such as better credit rating or a scholarship.

Motivating Examples

  1. Resume Score (UP): Based on the candidate's years of experience, best fit model (dotted purple line below) sometimes penalizes candidates for having more job experience. The model picks the age discrimination in the biased training samples.
    • Monotonicity shape constraints (blue line) guarantee that model uses job experience as positive evidence.

Such unfair responses are very much a danger in sparse regions of a feature space with modern over-parameterized nonlinear models.

  1. Law School Admissions (UP): Standard 2-DNN (and, gradient boosted tree) violate merit-based social norms and the "best-qualified" ethical principle by penalizing people for higher GPA or rewarding for lower GPA. (objectionable model)
    • Training a monotone calibrated linear model helps the cause, with similar test accuracy compared to DNN and GBT.

Non-uniform distribution of the training data means that some regions of the feature space were sparse and model may have overfit.

  1. Crimes and Misdemeanors:

    • (UP) Fines for misdemeanors would be fairer if they increase monotonically w.r.t. magnitude of illegality.
    • (FLF) Juvenile shouldn't penalized more heavily than an adult for the same crime, all else equal.
    • (FLF) Prefer to not give harsher penalties to first-time offenders than to repeat offenders.
  2. Pay:

    • (UP) Unfair to get paid less for doing more of the same work.
    • (UP) Desirable to workers if the recommended pay were constrained to be increasing monotonically as a function of the number of hours worked, all else equal.
    • (UP) Larger tips for expensive meals and higher tip percentages for more upscale establishments, in accordance with American societal norms.
  3. Medical Triage:

    • (FLF) More ethical or responsible to prioritize patients based on their risk or neediness.
    • (UP) Emergency room prioritization should treat patients that have waited longer first, if all other relevant characteristics are equal.

Contributions

  • Need for Monotonicity: Authors demonstrate that violation of monotonicity in ML models can result in significant ethical issues and risk of societal harm. These violations easily occur in practice with nonlinear ML models.
  • Existing (lattice) shape constrained ML approaches can ameliorate such problems.
  • In addition to consequentialist view of ML fairness, the authors introduce deontological aspect of fairness. They further provide analysis of the relation between proposed monotonicity constraints and statistical fairness notions.

Enforcing Monotonicity Constraints

Historically, monotonicity shape constraints have been used to capture prior knowledge and regularize estimation problems to improve a model's generalization to new test samples. Here, authors show that these constraints can and also should be used to ensure that ML models behave consistently with societal norms and prima facie duties.

However, it's likely that embedding monotonicity constraints to impose ethical principles may hurt test accuracy if the training and test data are biased. (performance drop was minimal for the experiments presented in this work)

Relation to Statistical Fairness

Most of the theoretical contribution has been presented in this sub-section: interaction between deontological monotonicity constraints and consequentialist statistical fairness goals which are based on aggregate outcomes.

  1. One-sided statistical parity (OSSP): Due to asymmetric goals such as FLF, OSSP requires that E[f(X, Z) | Z = j] <= E[f(X, Z) | Z = k] for any j<=k, where Z is the intended-monotone (protected) attribute. Though monotonicity can't guarantee OSSP, it implies a bound on the OSSP violation - maximum density ratio between the two groups (Z = j and Z = k).

    • Note: Converse doesn't hold.
    • Monotonic projection of f cannot have worse statistical parity violations on average. (Always doing better with monotone model)
  2. Bound for Binary Classifiers: Goal of statistical parity in this case is equivalent to marginal independence. So, the bound on statistical parity becomes a bound on the marginal probabilities.

  3. Equal opportunity: Less discussion has been provided on this in the paper. Read Lemma 4 for specific details of the bound.

Experiments

Setup

  • Calibrated linear model from TensorFlow Lattice 2.0 package was used in all the experiments.
  • Piece-wise calibration was done by dividing into K = 20 quantiles.
  • Joint training via projected SGD was further tuned by picking LR through grid search on powers of 10.

Credit Default (Taiwanese) : UP

Sourced from the UCI repository, the target is to predict whether a user defaulted on a payment in a time window, or not.

  • Unfair to many if they are penalized for paying their bills sooner, all else equal. Monotonicity constraints are included to abstain the model from penalizing early payments.
  • 1st experiment: Only 2 features: marital status and April repayment status (constrained).
  • 2nd experiment: All 24 features: monotonicity constraint on all 6 repayment features to avoid penalization.
  • Train and test accuracy were reported by the authors.

Monotonicity constraints have tiny effect on the performance of 1st experiment, but slightly worse for the 2nd one.

Since the train accuracy is not hurt, authors believe the lower test accuracy is simply due to the randomness of the sample.

Funding Proposals : FLF

Sourced from KDD Cup 2014, the target feature is a binary label that represents outcome of the project proposed by teachers - "exciting" (1) and "not exciting" (0).

  • Standard ML models won't favor the schools at higher poverty levels, or projects with greater impact.
  • Again, 2 experiments are conducted: first, poverty level and number of students impacted are considered. And, then all the features are included (same two constrained though).
  • Train and test AUC were reported by the authors, mostly because of the inherent balance (~5%).

In this case, monotonicity constraints to prefer poorer schools and greater student impact don't hurt the test AUC.

  • For comparison to statistical fairness, empirical values of OSSP and equal opportunity are shown below. Note that the monotonically constrained models do lower violations of both fairness goals. Improvement is smaller when model had all features than just 2 features.

Conclusion

  • Non-linear ML models can easily overfit noise or learn bias in a way that violates social norms or ethics about whether certain inputs should be allowed to negatively affect a score or decision.
  • This problem can be ameliorated by training with monotonicity constraints to reflect the desired principle.
  • The constraints can improve or bound (consequentialist) one-sided statistical fairness violations.
  • The constaints are fairly easy to explain, and reason about for laypeople.

Monotonicity constraints are a necessary and useful tool for creating responsible AI, but certainly not sufficient or application to all situations.

Open Questions

  • For the proposed method to be completely effective, all relevant features must be identified and constrained.
  • Special care could be taken to ascertain whether the non-sensitive information is highly correlated with the sensitive information (thereby, causing indirect discrimination).
  • This method is not directly applicable to unordered inputs such as addresses, photos, or voice signals.