12  Considerations for Responsible Data Science

Published

April 10, 2026

Keywords

responsible data science, data science life cycle, law, policy, ethics, ethcial frameworks

12.1 Introduction

12.1.1 Learning Outcomes

  • Differentiate legal, professional, and ethical considerations.
  • Identify and interpret professional Codes of Conduct.
  • Apply Philosophical Frameworks for ethical reasoning.
  • Identify sources of potential ethical issues in Data Science.
  • Practice responsible data science from legal, professional, and ethical perspectives.

12.1.2 References:

  • See references.

12.1.2.1 Other References

  • See links in the notes.
Warning

These notes provide general background to facilitate educational discussion.

They do not offer any form of legal advice or recommendations.

12.3 Ethical Frameworks

12.3.1 Some (of many) Ethical Frameworks 1

Consequentialist or Utilitarianism: greatest balance of good over harm (groups/individual).

  • Focus is on the consequences.
  • Choose the future outcomes that produce the most good.
  • Compromise is expected as the end justifies the means.

Duty or Deontology: Do your Duty , Respect the Rules, Be Fair, Follow Divine Guidance.

  • Focus is on the rules and whether the act violated the rules regardless of consequences.
  • Do what is “right” regardless of the consequences or emotions.
  • Everyone has the same duties and ethical obligations at all times.

Virtues: Live a Virtuous Life by developing the proper character traits.

  • Focus is on what society would say what a virtuous person would do.
  • Ethical behavior is whatever a virtuous person would do.
  • Tends to reinforce local cultural norms as the standard of ethical behavior.

Rights

  • Focus on the rights of the affected stakeholders, both individuals and groups.
  • Ethical behavior does not result in a violation of someone’s rights.
  • How do we know agree on is a “right” versus a “privilege” or “benefit”?

Frameworks can conflict with each other or are “wrong” in the extreme.

No single or simple right answer!

12.3.2 Using Frameworks in Ethical Decisions

  1. Recognize There May Be an Ethical Issue.
    • Assess the underlying definitions, facts, and assumptions and constraints
  2. Consider the Parties Involved, the stakeholders.
    • What individuals or groups might be harmed or benefit, and by when.
  3. Gather all Relevant Information. Are you missing key facts? Are they knowable?
  4. Formulate Actions and Evaluate under Alternative Frameworks.
    • What will produce the most good and do the least harm? ( Utilitarian)
    • What will be in accordance with existing rules or principles? (Duty)
    • What leads me to act as the sort of person I want to be? (Virtues)
    • What respects the rights of everyone affected by the decision? (Rights)
    • What treats people equally, equitably, or proportionately? ( Justice)
    • What serves the entire community, not just some members? (Common Good)
  5. Examine Alternatives and Make a Decision.
  6. Act and Observe.
  7. Assess and Reflect on the Outcomes.

If you don’t like the outcomes, restart the process.

12.4 Identifying Ethical Issues

12.4.1 Our Biased Brains Helped Us Survive

Our brains have evolved mechanisms to make quick decisions.

These are the source of “Unconscious Biases” or “Implicit Biases”.

  • “An unconscious association, belief, or attitude toward any social group” that can lead to stereotyping. (Kendry Cherry MsED 2025)
  • Under stress, we tend to bypass the higher-level cognitive centers that evolved later and take more time to reason.

Humans get comfortable with patterns which can lead to systematic deviations from making rational judgments.

Ethical Challenges can arise from our own implicit biases or the implicit biases of others affecting our data, thoughts, and actions.

Short-term choices under stress may be bad in the long run ….

12.4.2 Can Data and Algorithms Exhibit Harmful Bias?

The question goes back decades; modern AI systems are driving expanded research.

12.4.3 Three Articles for Consideration

  1. Higher error rates in classifying the gender of darker-skinned women than for lighter-skinned men (O’Brien 2019)

  2. Big Data used to generate unregulated e-scores in lieu of FICO scores for Credit in Lending] is on Canvas. (Bracey and Moeller 2018)

  3. Learning Analytics Can Violate Student Privacy (Raths 2018)

Discussion Questions

  • Is there an ethical issue or more than one? What is it?
  • Who is affected and who is responsible?
  • Pick one of the Professional Codes of Conduct or Guidelines. How would it apply?
  • What would you do differently or recommend?

12.4.4 More Examples of Ethical Issues

  • Contradictions and competition among legal, professional, and ethical guidelines.
  • Using biased data ( even unknowingly) or eliminating extreme values or small groups.
  • Using Proxies for “Protected” Attributes (even unknowingly).
  • Protection of Intellectual Property versus Explainability, Transparency and Accountability
  • Law of Unintended Consequences - people will use your products and solutions in “creative” ways that - May not align with your principles, or, - Be technically inappropriate.

So what can you do? What should you do?

12.5 Practicing Responsible Data Science.

What Can You Do? What Should You Do?

12.5.1 Consider Ethical Choices Across the DS Life Cycle

Responsible Data Science Life Cycle: Ask a question to Observe Outcomes

Ask a question: Equity or equality ? Stakeholders and Trade offs ? What are our interests? Recency or Confirmation Bias?

Frame the Analysis: What is the population? Role of proxy variables? How are metrics for “fairness”affecting groups/individuals? Do we need an IRB(APA 2022)?

Get Data: How was it collected? Was informed consent required/given? Is there balanced representation ? Selection Bias? Availability Bias? Survivorship Bias?

Shape Data: Are we aggregating distinct groups? How do we treat missing data? Are we separating training and testing data ?

Model and Analyze: How are we documenting assumptions, treating extreme values, or checking over-fitting? Are we checking multiple fairness and performance metrics?

Communicate Results: Are the graphs misleading? Did we cherry pick or data snoop? Are we reporting \(p\)-values and hyper-parameters?

Deploy/Implement: Is the deployment accessible to all?

Observe Outcomes: Can we check assumptions and analyze outcomes for bias?

Are we following professional guidelines from ASA, ACM, INFORMS, …?

12.5.2 Consider Frameworks for Responsible Data Science

Published by the Royal Statistical Society and the Institute and Faculty of Actuaries in A Guide for Ethical Data Science

  1. Start with clear user need and public benefit
  2. Be aware of relevant legislation and codes of practice
  3. Use data that is proportionate to the user need
  4. Understand the limitations of the data
  5. Ensure robust practices and work within your skill set
  6. Make your work transparent and be accountable
  7. Embed data use responsibly

(RSS-IFA 2021)

Department of Defense AI Capabilities shall be:

  • Responsible. … exercise appropriate levels of judgment and care, while remaining responsible for the development, deployment, and use….

  • Equitable. … take deliberate steps to minimize unintended bias in AI capabilities.

  • Traceable. … develop and deploy AI capabilities such that relevant personnel have an appropriate understanding …, including with transparent and auditable methodologies, data sources, and design procedures and documentation.

  • Reliable. … AI capabilities will have explicit, well-defined uses, and the safety, security, and effectiveness … will be subject to testing and assurance within those defined uses ….

  • Governable. … design and engineer AI capabilities to fulfill their intended functions while possessing the ability to detect and avoid unintended consequences, and the ability to disengage or deactivate deployed systems that demonstrate unintended behavior.

(DoD 2020)

AI Principles

  1. Bold innovation: We develop AI that assists, empowers, and inspires people in almost every field of human endeavor; drives economic progress; and improves lives, enables scientific breakthroughs, and helps address humanity’s biggest challenges. Avoid creating or reinforcing unfair bias. …
  2. Responsible development and deployment: Because we understand that AI, as a still-emerging transformative technology, poses evolving complexities and risks, we pursue AI responsibly throughout the AI development and deployment lifecycle, from design to testing to deployment to iteration, learning as AI advances and uses evolve. …
  3. Collaborative progress, together: We make tools that empower others to harness AI for individual and collective benefit. …

Feb 2025 Responsible AI Progress Report

(GoogleAI 2025)

IBM Principles for Trust and Transparency

  1. The purpose of AI is to augment human intelligence. The purpose of AI and cognitive systems developed and applied by IBM is to augment – not replace – human intelligence.
  2. Data and insights belong to their creator.
  • IBM clients’ data is their data, and their insights are their insights. Client data and the insights produced on IBM’s cloud or from IBM’s AI are owned by IBM’s clients. We believe that government data policies should be fair and equitable and prioritize openness.
  1. New technology, including AI systems, must be transparent and explainable.
  • For the public to trust AI, it must be transparent. Technology companies must be clear about who trains their AI systems, what data was used in that training and, most importantly, what went into their algorithm’s recommendations. If we are to use AI to help make important decisions, it must be explainable.

(IBM 2019)

A guide to Building “Trustworthy” Data Products

Based on the golden rule: Treat others’ data as you would have them treat your data

  • Consent - Get permission from the owners or subjects of the data before …

  • Clarity - Ensure permission is based on a clear understanding of the extent of your intended usage

  • Consistency - Build trust by ensuring third parties adhere to your standards/agreements

  • Control (and Transparency) - Respond to data subject requests for access/modification/deletion, e.g., the right to be forgotten

  • Consequences (and Harm) - Consider how your usage may affect others in society and potential unintended applications.

(Loukides, Mason, and Patil 2018)

The FAIR Guding Principles for scientific data management and stewardship aim to improve the “Findability, Accessibility, Interoperability, and Reuse” of digital assets.

  • Findable: Metadata and data should be easy to find for both humans and computers. Machine-readable metadata are essential for automatic discovery …
  • Accessible: Users need to know how data can be accessed,including authentication and authorization, e.g., make data downloadable from a credible repository …
  • Interoperable: Your data should be in non-proprietary, modern formats/software so it can be integrated with other data …
  • Reusable: Metadata and data should be well-described so they can be replicated and/or combined in different settings based on proper documentation and a clear and accessible data usage license.

(GO-FAIR n.d.)

12.5.3 To Be an Ethically Responsible Data Scientist …

Integrate ethical decision making into your environment.

As Davy Crockett might say, “Try to be sure you are right, then Go Ahead!”

12.5.4 Stay Current on Emerging Ideas

Try to stay on the fast moving train that is Responsible Data Science.

12.6 Considerations for Responsible Data Science Summary

You should now be able to better:

  • Differentiate legal, professional, and ethical considerations.
  • Identify and interpret professional Codes of Conduct.
  • Apply Philosophical Frameworks for ethical reasoning.
  • Identify sources of potential ethical issues in Data Science.
  • Practice responsible data science from legal, professional, and ethical perspectives.

You should also have greater competence in considering choices that practice and promote responsible data science today, and in the future.

Finally, when it comes to responsible Data Science, Jane Addams reminds us that thinking about ethics is not enough, …

“Action indeed is the sole medium of expression for ethics.”

AAAI ACM. n.d. “Conference on Artificial Intelligence, Ethics, and Society.” AAAI. Accessed January 12, 2026. https://aaai.org/conference/aies/.
ACM. 2021. “The Code for Computing Professionals.” https://www.acm.org/code-of-ethics.
———. n.d. ACM FAccT Conference.” Accessed January 12, 2026. https://facctconference.org/.
APA. 2022. FAQs about IRBs.” Https://Www.apa.org. https://www.apa.org/advocacy/research/defending-research/review-boards.
Areheart, Bradley A, and Jessica L Roberts. 2019. GINA, Big Data, and the Future of Employee Privacy.” The Yale Law Journal.
ASA. 2021. “Ethical Guidelines for Statistical Practice.” https://www.amstat.org/ASA/Your-Career/Ethical-Guidelines-for-Statistical-Practice.aspx.
Association, DS. 2021. “Code of Conduct.” https://www.datascienceassn.org/code-of-conduct.html.
AutoML. n.d. “Ethics and Accessibility Guidelines.” Accessed December 26, 2022. https://2023.automl.cc/ethics/.
Baer, Tobais. 2019. Understand, Manage, and Prevent Algorithmic Bias. 1st ed. Aprees.
Bracey, Kali, and Marguerite Moeller. 2018. “Legal Considerations When Using Big Data and Artificial Intelligence to Make Credit Decisions.” https://github.com/AU-datascience/data/blob/main/413-613/Bracey%20Moeller%20Unregulated%20e-scores%20March%202018.pdf.
BRITECITY. 2026. “Data Privacy Compliance Guide for Businesses GDPR, CCPA, State Laws BRITECITY.” https://britecity.com/articles/data-privacy-compliance-guide.
Brown University, Science and Technology Studies. 2021. “A Framework for Making Ethical Decisions.” https://www.brown.edu/academics/science-and-technology-studies/framework-making-ethical-decisions.
Bureau, US Census. 2023. “Understanding Differential Privacy.” Census.gov. https://www.census.gov/programs-surveys/decennial-census/decade/2020/planning-management/process/disclosure-avoidance/differential-privacy.html.
Calzon, Bernadita. 2021. “Misleading StatisticsReal World Examples For Misuse of Data.” BI Blog Data Visualization & Analytics Blog Datapine. https://www.datapine.com/blog/misleading-statistics-and-data/.
Commission, European. n.d. AI Act Enters into Force - European Commission.” Accessed January 9, 2025. https://commission.europa.eu/news/ai-act-enters-force-2024-08-01_en.
Desjardins, Jeff. 2017. “Every Single Cognitive Bias in One Infographic.” https://www.visualcapitalist.com/every-single-cognitive-bias/.
DHUD, US. 2021. “Fair Housing Act.” https://www.hud.gov/program_offices/fair_housing_equal_opp/fair_housing_act_overview.
Dilmegani, Cem. 2025. AI Ethics Dilemmas with Real Life Examples in 2026.” AIMultiple: High Tech Use Cases &Amp; Tools to Grow Your Business. https://research.aimultiple.com/ai-ethics/.
DoD, US. 2020. DOD Adopts 5 Principles of Artificial Intelligence Ethics.” https://www.defense.gov/Explore/News/Article/Article/2094085/dod-adopts-5-principles-of-artificial-intelligence-ethics/.
DOEd, US. n.d. “Family Educational Rights and Privacy Act (FERPA).” Guides. Accessed April 1, 2020. https://studentprivacy.ed.gov/ferpa.
Edelman, Gilead. 2020. “Everything You Need to Know About the CCPA.” https://www.wired.com/story/ccpa-guide-california-privacy-law-takes-effect/.
EEOC, US. 2021a. “Genetic Information Nondiscrimination Act of 2008.” https://www.eeoc.gov/laws/statutes/gina.cfm.
———. 2021b. “Laws Enforced by EEOC.” https://www.eeoc.gov/statutes/laws-enforced-eeoc.
ExpertPanel, Forbes. 2024. “20 Tips For Addressing Unconscious Bias At Work From The Top Down.” Forbes. https://www.forbes.com/councils/forbeshumanresourcescouncil/2024/09/13/20-tips-for-addressing-unconscious-bias-in-the-workplace-starting-from-the-top-down/.
Fleisher, Will. 2024. AI Ethics.” Center for Digital Ethics. https://digitalethics.georgetown.edu/research-programs/artificial-intelligence-ai-ethics/.
FTC. 2013a. “Children’s Online Privacy Protection Rule ("COPPA").” https://www.ftc.gov/enforcement/rules/rulemaking-regulatory-reform-proceedings/childrens-online-privacy-protection-rule.
FTC, US. 2013b. “Fair Credit Reporting Act.” https://www.ftc.gov/enforcement/statutes/fair-credit-reporting-act.
———. 2018. “Credit Reporting Information.” https://www.ftc.gov/news-events/media-resources/consumer-finance/credit-reporting.
GO-FAIR. n.d. FAIR Principles.” GO FAIR. Accessed January 11, 2026. https://www.go-fair.org/fair-principles/.
GoogleAI. 2025. “Our Principles.” Google AI. https://ai.google/principles/.
HIPAA Partners Team. 2025. HIPAA Compliant Data Analytics: Privacy-Preserving Healthcare Insights.” https://hipaapartners.com/blog/hipaa-compliant-data-analytics-privacy-preserving-healthcare-insights.
IBM. 2019. IBMS Principles for Data Trust and Transparency.” IBM Policy. https://www.ibm.com/policy/trust-principles/.
IEAI-ML. 2021. “Awesome AI Guidelines.” https://github.com/EthicalML/awesome-artificial-intelligence-guidelines.
Implicit, Project. 2022. “Testing.” https://implicit.harvard.edu/implicit/.
INFORMS. 2021. INFORMS Ethics Guidelines.” https://www.informs.org/About-INFORMS/Governance/INFORMS-Ethics-Guidelines.
Japkowicz, Nathalie, and Zois Boukouvalas. 2024. Machine Learning Evaluation Pattern Recognition and Machine Learning. Cambridge University Press. https://www.cambridge.org/us/academic/subjects/computer-science/pattern-recognition-and-machine-learning/machine-learning-evaluation-towards-reliable-and-responsible-ai, https://www.cambridge.org/us/academic/subjects/computer-science/pattern-recognition-and-machine-learning.
Kendry Cherry MsED. 2025. “How Does Implicit Bias Influence Behavior?” Verywell Mind. https://www.verywellmind.com/implicit-bias-overview-4178401.
Laurie Harris. 2025. “Regulating Artificial Intelligence: U.S. And International Approaches and Considerations for Congress- a CRS Report.” Legislation. https://www.congress.gov/crs-product/R48555.
Law and More Attorneys. 2025. “Data Privacy In 2025: How The GDPR Is Evolving With AI & Big Data.” https://highpowerlasertherapy.com/law/blog/data-privacy-in-2025-how-the-gdpr-is-evolving-with-ai-and-big-data/.
Li, Jingyang, and Guoqiang Li. 2025. “Triangular Trade-off Between Robustness, Accuracy, and Fairness in Deep Neural Networks: A Survey.” ACM Comput. Surv. 57 (6): 140:1–40. https://doi.org/10.1145/3645088.
Loukides, Mike, Hilary Mason, and DJ Patil. 2018. Ethics and Data Science. 1st ed. O’Reilly Media, Inc. https://www.oreilly.com/library/view/ethics-and-data/9781492043898/.
Merriam-Webster. 2021. “Next Stop: Trolley Problem.” https://www.merriam-webster.com/words-at-play/trolley-problem-moral-philosophy-ethics.
Nature. 2025. “Let 2026 Be the Year the World Comes Together for AI Safety.” Nature 649 (8095): 7–7. https://doi.org/10.1038/d41586-025-04106-0.
O’Brien, Matt. 2019. MIT Researcher Exposing Bias in Facial Recognition Tech ...” https://www.insurancejournal.com/news/national/2019/04/08/523153.htm.
O’Neil, Cathy. 2017. Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. https://weaponsofmathdestructionbook.com/.
Obermeyer, Ziad, Rebecca Nissan, Michael Stern, Stehpanie Eaneff, Emily Joy Bembeneck, and Sendhil Mullainathan. 2021. “Algorithmic Bias Playbook.” https://www.chicagobooth.edu/-/media/project/chicago-booth/centers/caai/docs/algorithmic-bias-playbook-june-2021.pdf.
Psychology, Practical. 2017. “12 Cognitive Biases Explained -.” https://www.youtube.com/watch?v=wEwGBIr_RIw.
Radanliev, Petar. 2025. AI Ethics: Integrating Transparency, Fairness, and Privacy in AI Development.” Applied Artificial Intelligence 39 (1): 2463722. https://doi.org/10.1080/08839514.2025.2463722.
Raths, David. 2018. “When Learning Analytics Violate Student Privacy.” https://campustechnology.com/articles/2018/05/02/when-learning-analytics-violate-student-privacy.aspx.
Rismani, Shalaleh, Renee Shelby, Leah Davis, Negar Rostamzadeh, and Ajung Moon. 2025. “Measuring What Matters: Connecting AI Ethics Evaluations to System Attributes, Hazards, and Harms.” arXiv. https://doi.org/10.48550/arXiv.2510.10339.
RSS-IFA. 2021. “Data Science Ethics Guidelines UK.” https://www.actuaries.org.uk/upholding-standards/data-science-ethics.
Stanford Human Centered AI. 2025. “The 2025 AI Index Report.” https://hai.stanford.edu/ai-index/2025-ai-index-report.
Susarla, Anjana. 2025. “How States Are Placing Guardrails Around AI in the Absence of Strong Federal Regulation.” The Conversation. https://doi.org/10.64628/AAI.ysd4mntvt.
Us Dept of Labor. 2024. “Guidance on the Protection of Personal Identifiable Information.” DOL. http://www.dol.gov/general/ppii.
US DHHS, Office for Civil. 2015. HIPAA for Professionals.” Text. https://www.hhs.gov/hipaa/for-professionals/index.html.
Vattikuti, Manoj Chowdary. 2025. “7 Cognitive Biases That Affect Your Data Analysis (and How to Overcome Them).” Open Association of Research Society, United States. https://associationofresearch.org/7-cognitive-biases-that-affect-your-data-analysis-and-how-to-overcome-them/.
Vigen, Tyler. 2021. “15 Insane Things That Correlate With Each Other.” http://tylervigen.com/spurious-correlations.
Weissgerber Tracey L., Tracey et al. 2019. “Reveal, Don’t Conceal.” Circulation 140 (18): 1506–18. https://doi.org/10.1161/CIRCULATIONAHA.118.037777.
Wikipedia. n.d. “General Data Protection Regulation.” Wikipedia. Accessed December 26, 2022. https://en.wikipedia.org/w/index.php?title=General_Data_Protection_Regulation&oldid=1128013242.

  1. Brown University (2021)↩︎