This summary contains everything from the lectures that you need to know for the endterm exam of RDS. It is made specifically for RDS in 2024. For the summary, I have used my own lecture notes and the lecture slides.
Summary Responsible Data Science
(INFOB3RDS) Lecture 1-6 Endterm
Glossary
Act utilitarianism: choose action that maximizes utility
Algorithmic bias: bias due to wrong algorithm, does not work well with your data.
Ambiguity effect (cognitive bias): always prefer known risks over unknown risks
Analytical bias (data bias): results from how results are evaluated
Anchoring effect (cognitive bias): people rely too much on first piece of information (which
can be irrelevant)
Appropriate details (XAI property): (the amount of) details in explanation should be relevant
Association bias: data reinforces cultural bias, (mostly ‘gender bias’), result of training data.
Attraction effect (decoy effect): chooser tends to choose items that have more (slightly
worse) competitors, instead of just considering logical alternatives (choice is influenced)
Automation bias (human bias: blindly trusting technology. Very psychological.
Backward chaining: explain results, get facts from results
Bandwagon effect (cognitive bias): basing your decision on what the crowd does
Base rate fallacy (cognitive bias): people ignore statistical base rates of incidents
Black-box algorithm: model representation shaped by algorithm, not understandable to
humans.
Categorical imperatives: things you should follow regardless your desires
Clarity (XAI property): explanations should be clearly understandable
Cognitive bias: tricks our thinking instead of vision.
Completeness (XAI property): explanations should explain everything you need to know
Confirmation bias (cognitive bias): we tend to seek information confirming our beliefs
Confirmation bias (human bias): “see what you want to see”, results of interpretations
Contractarianism: we make up morality with an agreed-upon contract, by cooperation.
Decoy effect: (compare to attraction effect): 3rd (inferior alternative) changes preference
Demographic parity (algorithmic fairness measure): low difference between proportion of
privileged subjects in both groups.
Dimensions of variation (XAI property): explanation should reveal where boundaries are
(what is needed to change outcome)
Disparate impact (algorithmic fairness measure): ratio of certain group in privileged &
unprivileged groups should be equal
Dunning-kruger effect (cognitive bias): low-ability people overestimate themselves, while
high-ability people underestimate themselves
Equalized odds (algorithmic fairness measure): low difference between TPR & FPR of both
groups
Equal opportunity (algorithmic fairness measure): TPR (almost) equal groups are treated
similarly.
Exclusion bias (human bias): systematically excluding certain data
Feature relevance: measure relevance of each feature. Loal: feature contributions for
specific prediction; Global: take all predictions into account (explain entire model overall)
Formula of humanity: humans are the end, not the means.
Forward chaining: derive results from facts
, Framing effect (cognitive bias): caused by perspective
Fundamental attribution error (cognitive bias): credit success to yourself but failure to others
Funding bias (human bias): favor interest group / sponsors in study
General (“strong”) AI: AI that can learn and use intelligence autonomously.
Hasty generalization (cognitive bias): generalize too fast, while your experiences are limited
Hindsight bias (cognitive bias): saying “i knew it all along” afterwards
Human bias: human psychological disposition to oversee things.
Hypothetical imperatives: if you desire X, you should do Y
Identifiable victim effect (cognitive bias): people ignore impersonal representations of e.g.
poverty, but will help a specific person
Inertia (status quo bias) (cognitive bias): stick to current situation, fearing change
Jens Gulden: main lecturer of RDS.
Kantianism: morality is constant. morality depends on intentions, not on outcome.
Layer-wise Relevance Propagation (LRP): understand ANN in the background, by
propagating the prediction backwards. Understand which parameters have highest influence.
Local explanations: model agnostic explanation: does not care about internal model, but only
about a specific input instance (local)
Measurement bias (data bias): data inaccurately reflects population, badly recorded.
Metaethics: foundation / origins of ethical principles. Grounding problem: looking for clear
unmoving moral belief.
Model simplification (XAI): simplifying a model, e.g. constructing a DT from ANN.
Moral absolutism (in moral realism: most extreme version of moral realism, moral facts are
unchangeable and universal.
Moral anti-realism (metaethical belief): there are no objective moral facts, they are
expressions of our emotional attitude rather than of the state of the world.
Moral realism (metaethical belief): objective morals exist, independent of human beliefs.
Moral relativism (in moral realism): moral facts are true / false only relative to specific
contexts
Moral subjectivism (in moral anti-realism): morals can be true or false, it is a matter of
preference
Narrow (“weak”) AI: statistically trained AI, works for predefined problems
Observability (XAI property): explanation should explain internal mechanism
Observer bias (human bias): researcher projects his expectations in the research (in
decisions)
Omitted variable bias (data bias): data lacks relevant features
Outgroup homogeneity bias (cognitive bias): people out of your own group seem more alike
Overgeneralization (human bias): generalize too quickly (I have seen 10 swans and they
were all white, therefore black swans do not exist)
Perceptual bias: distorts our perceptions / sensory experience.
Pseudonymization by encryption: pseudonymize data by encrypting sensitive data.
Pseudonymization by tokenization: pseudonymize data by giving each entry an anonymous
ID (token) based on name
Racial / gender bias: resulting from selection bias of training data.
Recall bias: Similar types are labeled consistently, due to human interpretation differences
Rule utilitarianism: maximize utility in the long run
Selection bias (data bias): some data points have more chance to be included, no proper
randomization is used
Soundness (XAI property): explanations should be consistent, reasonable, plausible
The benefits of buying summaries with Stuvia:
Guaranteed quality through customer reviews
Stuvia customers have reviewed more than 700,000 summaries. This how you know that you are buying the best documents.
Quick and easy check-out
You can quickly pay through credit card or Stuvia-credit for the summaries. There is no membership needed.
Focus on what matters
Your fellow students write the study notes themselves, which is why the documents are always reliable and up-to-date. This ensures you quickly get to the core!
Frequently asked questions
What do I get when I buy this document?
You get a PDF, available immediately after your purchase. The purchased document is accessible anytime, anywhere and indefinitely through your profile.
Satisfaction guarantee: how does it work?
Our satisfaction guarantee ensures that you always find a study document that suits you well. You fill out a form, and our customer service team takes care of the rest.
Who am I buying these notes from?
Stuvia is a marketplace, so you are not buying this document from us, but from seller danielgeelhoed. Stuvia facilitates payment to the seller.
Will I be stuck with a subscription?
No, you only buy these notes for $6.29. You're not tied to anything after your purchase.