Data Ethics: A Checklist with 7 Points to Consider
Data Ethics: A Checklist with 7 Points to Consider

Data Ethics: A Checklist with 7 Points to Consider

Data ethics involves defining and promoting ethical principles for handling personal data; This includes responsible practices for collecting, using, storing, sharing, and disposing of data. Data ethics aims to ensure that data is managed morally, soundly, and responsibly.

In an increasingly data-driven world, where vast amounts of personal information are being collected and processed, data ethics seeks to address the ethical challenges and implications that arise. It recognizes the importance of respecting individuals' privacy, autonomy, and rights in data handling.

Data ethics involves informed consent, transparency, fairness, accountability, data minimization, purpose limitation, data security, and the responsible use of emerging technologies like artificial intelligence and machine learning. It also examines the potential biases and discrimination resulting from data collection and analysis.

It is recommended that both organizations and individuals adopt ethical frameworks and practices to handle data that uphold individual rights and promote trust, integrity, and social responsibility; This can be achieved through privacy policies, privacy impact assessments, data anonymization or pseudonymization, and promoting data literacy and awareness among stakeholders.

Our focus should be on maintaining data ethics to balance maximizing the advantages of data-driven technologies while preserving individual privacy and well-being. We are constantly working towards setting ethical norms and standards that promote responsible data use, fostering a more reliable and ethical data ecosystem.

Below are 7 points to consider.

1. Start with precise user needs and public benefit

Describe the user's need:

  • Does everyone in the team understand the user's needs?
  • How does this benefit the public?
  • What would be the harm in not using data science — what might not be met?
  • Do you have supporting evidence for the approach being likely to meet a user's need or provide public benefit?

2. Be aware of relevant legislation and codes of practice

List the pieces of legislation, codes of practice, and guidance that apply to your project:

  • Do all team members understand how relevant laws applicable to the project?
  • If necessary, have you consulted with relevant experts?
  • Could you let me know if you've been able to speak with your information assurance team?
  • If using personal data, do you understand obligations under data protection legislation?

3. Use data that is proportionate to the user's need

If using personal data, have you answered the questions for determining proportionality?

  • You'll need to include evidence to support any decision.
  • If using personal data, what measures are in place to control access?
  • How widely are you searching personal data?
  • How can you meet the project aim using the minimum personal data possible?
  • Is there a way to achieve the same aim with less identifiable data?
  • Can you use synthetic data?
  • Has the data being used been provided for your analysis?
  • Using data that the public has freely volunteered, would your project jeopardize people providing this again in the future?
  • Could you clearly explain why you need to use that data for members of the public?
  • Is there a fair balance between the rights of individuals and the interests of the community?

4. Understand the limitations of the data

Identify the potential limitations of the data source(s) and how they are being mitigated:

  • What data source(s) is being used?
  • Are all metadata and field names clearly understood?
  • What processes do you have in place to ensure and maintain data integrity?
  • Is there a plan in place to identify errors and biases?
  • What are the caveats?

5. Use robust practices and work within your skillset

Explain the relevant expertise and approaches that are being employed to maximize the efficacy of the project:

  • Describe the disciplines involved and why.
  • Is there expertise the project requires that you don’t currently have?
  • Have you designed the approach with a policy team or subject matter expert(s)?
  • Has all subject matter context, from policy experts or otherwise, been considered when determining the appropriate loss function for the model?
  • If necessary, how can you (or with external scrutiny) check that the algorithm makes the right output decision when new data is added?
  • How has reproducibility been made? Could another analyst repeat your procedure based on your documentation?
  • How confident are you that the algorithm is robust, and that any assumptions are met?
  • What is the quality of the model outputs, and how does this stack up against the project objectives?
  • If using data about people, is it possible that a data science technique is basing analysis on proxies for protected variables which could lead to a discriminatory policy decision?
  • What processes are in place to ensure and maintain data integrity?

6. Make your work transparent and be accountable

Describe how you have considered making your work transparent and accountable:

  • Have you been able to speak with your organization to find out if you can speak about your project openly?
  • Have you considered how internal and external engagement could benefit your project?
  • How interpretable are the outputs of your work?
  • Could you explain how approaches were designed in plain English to other practitioners, policymakers, and, if appropriate, the public?
  • Can you openly publish your methodology, metadata about your model, and the model itself, e.g., on Github?
  • Can you get peers to review your Pull Requests?

7. Embed data use responsibly

Describe the steps taken to ensure any insight is managed responsibly:

  • How many people will be affected by the new model, insight, or service?
  • Who are the insight, model, or new service users?
  • Do users have the appropriate support and training to maintain the new technology?
  • Do you know if future events have been planned?
  • Is your implementation plan correlated with the impact of a particular model?
  • Please let me know how often you'll report on these plans to senior reporting officers.


Murat Durmus

(Author of the Book: "MINDFUL AI: Reflections on Artificial Intelligence")

Irina Raicu

Director of the Internet Ethics Program at the Markkula Center for Applied Ethics, Santa Clara University

2y

For those looking for teaching resources on this: "An Introduction to Data Ethics": https://www.scu.edu/ethics/focus-areas/technology-ethics/resources/an-introduction-to-data-ethics/

Like
Reply
Varun Madiyal

I help insurers to build digital & data driven solutions | Analytics & Insights | ML & AI | HealthTech & InsureTech | Speaker & Author | Thought Leadership & Mentoring |

2y

Insightful read 💗

To view or add a comment, sign in

More articles by Murat Durmus

  • So Many AIs, So Little Time

    "So many AIs, so little time" – a phrase that, on the surface, seems drowned in the desire for the new. But upon…

  • The Time Will Come When We Will Envy Sisyphus

    Camus would have been tragically amused that much is certain. Sisyphus, condemned to roll his boulder up the mountain…

    5 Comments
  • A Machine that Calls You by Your First Name

    Once protected by silence and solitude, the modern soul now bleeds metadata. There was a time when solitude was a…

  • Debugging the Wrong System

    We work feverishly to prune the branch of artificial intelligence, developing ever finer algorithms to mimic the wisdom…

    2 Comments
  • Embodied Meaning

    To transcend language, AI must engage with what language only hints at: context, sensation, embodiment, consciousness…

    6 Comments
  • Fragments of the Furnace

    Humanity no longer stands at the center of creation, but awkwardly beside it, like a former rock star watching his…

    2 Comments
  • The Fog of Synthetic Souls

    What happens when you can no longer distinguish AI from humans (in many instances, that's already the case), we lose…

    1 Comment
  • Checkbox Faith

    Trust used to be felt in the spaces between words. In eye contact, in silence, in someone handing you the bigger half…

    1 Comment
  • The Illusion of Control or Chainsaws and Choices

    The greatest danger of AI is that we give it power while acting as if it were still a tool. This is like giving a…

    2 Comments
  • Bias is a Mirror, Not a Bug

    In pursuing justice, perfection is an illusion, while participation is a moral imperative. Fairness requires more than…

    1 Comment

Insights from the community

Others also viewed

Explore topics