Appendix A: Research Methods for Data Ethics

This appendix provides a concise primer on the research methods most commonly used to study data governance, privacy, algorithmic fairness, and related topics in data ethics. Whether you are reading studies cited in this textbook, designing your own research for a capstone project, or evaluating claims made by companies and policymakers, understanding these methods will sharpen your analytical judgment.


A.1 Why Research Methods Matter for Data Ethics

Data ethics is an inherently interdisciplinary field. Its questions -- Who is harmed by algorithmic systems? Does consent actually protect users? Do data governance frameworks reduce harm in practice? -- require evidence, not just argument. The studies referenced throughout this textbook use a range of methods drawn from computer science, social science, law, and philosophy. Understanding those methods is essential for three reasons:

  1. Evaluating claims. When a company claims its algorithm is "fair," or a study finds that privacy policies are "unreadable," you need to know how those conclusions were reached and whether the methods support them.

  2. Designing your own research. The capstone projects in Chapter 40 ask you to conduct audits, design governance frameworks, and write policy briefs. Each of these requires systematic inquiry.

  3. Engaging with the literature. As data ethics evolves rapidly, staying current requires reading new studies and assessing their quality independently.


A.2 Qualitative Methods

Qualitative research seeks to understand phenomena through in-depth exploration of experiences, meanings, and contexts. It produces rich, detailed data that captures complexity and nuance.

A.2.1 Interviews

What they are: Structured, semi-structured, or unstructured conversations with research participants designed to elicit their perspectives, experiences, and interpretations.

When to use them: When you need to understand how and why people think, feel, or behave in particular ways. In data ethics, interviews are valuable for understanding how privacy professionals navigate organizational pressures (as Ray Zhao's accounts illustrate throughout this textbook), how affected communities experience algorithmic decision-making, or how policymakers balance competing interests.

Key considerations: - Sampling -- Who is interviewed matters enormously. Interviewing only corporate executives about data governance produces a different picture than interviewing data subjects, community organizers, or frontline workers. - Semi-structured formats are most common in data ethics research: the researcher has a set of guiding questions but allows the conversation to follow emergent themes. This balances consistency across interviews with the flexibility to explore unexpected insights. - Positionality -- The researcher's identity and institutional affiliation shape what participants are willing to share. A researcher affiliated with a technology company will elicit different responses than one affiliated with a civil liberties organization.

Example from this textbook: The chapter on algorithmic management (Chapter 33) draws on interview-based studies where gig workers described their experiences with opaque rating systems and automated deactivation.

A.2.2 Focus Groups

What they are: Guided group discussions (typically 6-10 participants) led by a moderator, designed to explore shared and divergent perspectives on a topic.

When to use them: When group dynamics and social negotiation of meaning are important. Focus groups are particularly useful for understanding how communities collectively make sense of data governance issues -- for example, how residents of a neighborhood discuss the installation of surveillance cameras.

Key considerations: Group dynamics can suppress minority viewpoints. A skilled moderator must ensure that dominant voices do not crowd out dissenting or marginal perspectives.

A.2.3 Ethnography and Participant Observation

What they are: Extended immersion in a social setting, where the researcher observes (and sometimes participates in) the activities, interactions, and practices of a community or organization.

When to use them: When you need to understand how data practices operate in context, beyond what people say they do. The gap between stated policy and actual practice is a recurring theme in this textbook, and ethnography is the primary method for documenting it.

Example: Studies of content moderation practices at social media companies (referenced in Chapter 31) have used ethnographic methods to reveal the emotional toll on moderators and the gap between platform policies and enforcement realities.

A.2.4 Content Analysis

What they are: Systematic analysis of texts, documents, images, or media to identify patterns, themes, and meanings.

When to use them: For analyzing privacy policies, terms of service, corporate ethics statements, regulatory documents, or media coverage of data governance issues. Content analysis can be qualitative (thematic analysis of what privacy policies say) or quantitative (counting the frequency of specific terms across a corpus of policies).

Example from this textbook: The analysis of consent forms in Chapter 9 draws on content analyses that measured readability, length, and the specificity of data use descriptions across hundreds of privacy policies.


A.3 Quantitative Methods

Quantitative research seeks to measure, count, and statistically analyze phenomena. It produces numerical data that can be generalized to larger populations when properly designed.

A.3.1 Surveys

What they are: Standardized questionnaires administered to a sample of respondents, designed to measure attitudes, beliefs, behaviors, or experiences.

When to use them: When you need to measure the prevalence or distribution of a phenomenon across a population. Surveys are widely used in data ethics research to measure privacy attitudes (the "privacy paradox" research in Chapter 11), trust in algorithmic systems, awareness of data rights, or support for regulatory interventions.

Key considerations: - Sampling bias -- Online surveys systematically exclude populations without internet access, which is precisely the population most vulnerable to digital exclusion (Chapter 32). - Question wording -- Small changes in phrasing can dramatically alter responses. The difference between "Do you trust companies to protect your data?" and "Do you believe companies adequately protect your data?" can produce different results. - Social desirability bias -- Respondents may overstate their privacy concern or understate their willingness to trade privacy for convenience. - Cross-cultural validity -- Survey instruments developed in Western, English-speaking contexts may not translate meaningfully across cultures (Chapter 37).

A.3.2 Experimental Design

What they are: Studies that manipulate one or more variables (independent variables) while controlling others to measure the effect on an outcome (dependent variable). Experiments can be conducted in laboratories, in the field, or online.

When to use them: When you need to establish causal relationships. While surveys and observational studies can identify correlations, experiments are the gold standard for determining whether one factor actually causes another.

Example from this textbook: The Vosoughi, Roy, and Aral (2018) study on the spread of false news on Twitter (Chapter 31) used observational methods, not experiments, which means it could document what spreads faster but not definitively why. In contrast, experimental studies of prebunking (Chapter 31) randomly assigned participants to receive inoculation messages or not, allowing causal claims about effectiveness.

Key considerations: - A/B testing -- A common form of online experimentation where users are randomly assigned to different versions of a product feature. The ethical concerns of A/B testing without informed consent are discussed in Chapter 9. - Field experiments -- Conducted in real-world settings, these have high external validity but raise ethical concerns about experimenting on people without their knowledge. - Randomized controlled trials (RCTs) -- The strongest experimental design, widely used in medicine and increasingly applied to policy evaluation. In data governance, RCTs have been used to test the effectiveness of different privacy notice formats.

A.3.3 Algorithmic Auditing

What they are: Systematic testing of algorithmic systems to assess their behavior, particularly for bias, accuracy, and fairness. Audits can be internal (conducted by the organization deploying the system) or external (conducted by researchers, journalists, or regulators).

When to use them: When you need to evaluate whether an algorithmic system produces discriminatory or harmful outcomes. The bias auditing methods discussed in Chapters 14 and 15 are research methods as well as governance tools.

Key considerations: - Access -- External auditors often lack access to the algorithm's code, training data, or internal documentation. Many audits therefore rely on black-box testing: sending inputs and analyzing outputs without knowledge of the system's internals. - Benchmark selection -- The dataset used to audit a system significantly affects the findings. The Gender Shades study (Buolamwini and Gebru, 2018) produced its groundbreaking results in part because it used a benchmark that was more demographically representative than existing benchmarks. - Reproducibility -- Algorithmic systems change over time (model updates, data refreshes), making audit results a snapshot rather than a permanent characterization.

A.3.4 Statistical Analysis of Observational Data

What they are: Analysis of naturally occurring data (not generated through experiments) using statistical methods to identify patterns, correlations, and associations.

When to use them: When experimental manipulation is impractical or unethical. Much data ethics research analyzes existing datasets -- court records (COMPAS analysis), hiring data (Amazon hiring algorithm), healthcare claims (Obermeyer health algorithm bias) -- to identify patterns of discrimination or inequity.

Key limitations: Observational data cannot establish causation. Correlations may be driven by confounding variables that the analysis does not account for. This is why the chapter on bias (Chapter 14) emphasizes the distinction between correlation and causation in interpreting algorithmic outputs.


A.4 Case Study Research

What it is: In-depth investigation of a specific instance, event, organization, or phenomenon, using multiple data sources (documents, interviews, observation, quantitative data) to build a comprehensive understanding.

When to use it: When the phenomenon is complex, context-dependent, and not easily separated from its environment. Case studies are the dominant method in data governance research because governance challenges are deeply embedded in organizational, legal, and cultural contexts.

Example from this textbook: The VitraMed thread is a longitudinal case study that traces how a single organization's data ethics challenges evolve over time. The Detroit smart city thread is a community-level case study. Each chapter's case studies (the two case study files accompanying each chapter) use the case study method.

Key considerations: - Generalizability -- A single case study cannot prove that its findings apply to all similar situations. However, case studies can generate hypotheses, illustrate mechanisms, and reveal dynamics that quantitative studies might miss. - Triangulation -- Strong case studies draw on multiple data sources and methods, allowing the researcher to cross-check findings and build a more robust account. - Researcher judgment -- Case study research requires extensive interpretation, making the researcher's analytical framework and potential biases particularly important to acknowledge.


A.5 Mixed Methods

What they are: Research designs that combine qualitative and quantitative methods within a single study or program of research.

When to use them: When neither qualitative nor quantitative methods alone can answer the research question. In data ethics, mixed methods are common because the field requires both measurement (how large is the bias? how many people are affected?) and understanding (what does the bias mean to those affected? how do organizations respond to audit findings?).

Example: The Obermeyer et al. (2019) study on racial bias in healthcare algorithms (Chapter 14) combined quantitative analysis of algorithmic outputs with qualitative investigation of why healthcare spending was used as a proxy for health need. The quantitative analysis identified the bias; the qualitative investigation explained its structural roots.


A.6 Evaluating Research Quality

Not all research is equally trustworthy. When reading studies cited in this textbook or encountered in your own research, evaluate them against the following criteria:

A.6.1 Validity

  • Internal validity: Does the study's design support the causal claims it makes? A study that claims A causes B must rule out alternative explanations (confounding variables, selection bias, measurement error).
  • External validity (generalizability): Do the findings apply beyond the specific sample or context studied? A study of privacy attitudes among US college students may not generalize to other populations, cultures, or age groups.
  • Construct validity: Does the study actually measure what it claims to measure? When a study claims to measure "privacy concern," does its survey instrument capture genuine concern or merely stated preference?

A.6.2 Reliability

Can the study's findings be reproduced? If another researcher used the same methods and data, would they reach the same conclusions? Algorithmic audits face particular reliability challenges because the systems they audit change over time.

A.6.3 Transparency

Does the study disclose its methods, data, analytical procedures, and limitations clearly enough for critical evaluation? Studies that do not share their data or code are harder to evaluate and reproduce.

A.6.4 Ethical Conduct

Was the research conducted ethically? Did it obtain informed consent from participants? Did it protect confidential information? Was it reviewed by an institutional ethics body? These questions are not merely procedural -- they reflect the same values that the research itself may be studying.

A.6.5 Funding and Conflict of Interest

Who funded the research? Does the funding source create potential conflicts of interest? A study of algorithmic fairness funded by the company whose algorithm is being evaluated should be scrutinized more carefully than one funded by an independent foundation -- not because corporate-funded research is inherently biased, but because the potential for bias exists and should be acknowledged.


Research about data ethics must itself be conducted ethically. This creates a recursive obligation: the methods used to study privacy, consent, and power must respect the very principles they investigate.

A.7.1 Institutional Review Boards (IRBs) and Ethics Committees

In the United States, any research involving human subjects conducted at an institution receiving federal funding must be reviewed by an Institutional Review Board (IRB). Equivalent bodies exist in other countries: Research Ethics Committees (UK), Human Research Ethics Committees (Australia), and National Ethics Committees (EU member states).

IRBs evaluate research proposals for: - Risk-benefit balance: Do the potential benefits of the research justify the risks to participants? - Informed consent: Are participants adequately informed about the study's purpose, procedures, risks, and their right to withdraw? - Privacy and confidentiality: Are participants' data adequately protected? - Vulnerable populations: Are special protections in place for research involving children, prisoners, people with cognitive impairments, or other vulnerable groups?

Limitations of IRBs for data ethics research: - IRBs were designed primarily for biomedical and behavioral research and may lack expertise in evaluating computational research involving large datasets, algorithmic systems, or online platforms. - Research involving publicly available data (social media posts, government records) may not require IRB review under current regulations, even when it involves identifiable individuals. - Corporate research (A/B testing, product experimentation) is generally not subject to IRB oversight because it is conducted by private companies, not federally funded institutions. This creates a significant governance gap, as discussed in Chapter 9.

Informed consent in research requires that participants: 1. Understand the purpose, procedures, and duration of the study 2. Understand the risks and potential benefits of participation 3. Know that participation is voluntary and can be withdrawn at any time 4. Understand how their data will be used, stored, and protected

Challenges specific to data ethics research: - Research on social media: Can researchers ethically analyze public social media posts without the users' consent? The posts are public, but the users did not anticipate their content would be studied. - Research on algorithmic systems: Black-box auditing may require creating fake accounts or submitting test applications, which may violate platforms' terms of service. Is this ethically justified when it serves the public interest? - Research involving existing datasets: When researchers analyze datasets collected by others (e.g., analyzing COMPAS data released by ProPublica), the individuals in the dataset did not consent to be research subjects. What obligations does this create?

A.7.3 Power Dynamics in Research

Research about marginalized communities -- communities that are disproportionately affected by surveillance, algorithmic bias, and data extraction -- raises particular ethical concerns:

  • Extractive research: Research that collects data from communities, produces academic publications, and provides no tangible benefit to the community replicates the extractive dynamics it may be studying. The CARE Principles for indigenous data governance (Chapter 32) provide a framework for non-extractive research.
  • Participatory research: Approaches such as community-based participatory research (CBPR) involve community members as co-researchers rather than subjects, addressing power imbalances in the research process.
  • Reciprocity: Ethical research should provide something of value to the communities it studies -- whether through shared findings, capacity building, policy advocacy, or direct material support.

A.8 Choosing a Method for Your Research

When designing research for capstone projects or independent study, consider the following decision framework:

Question Suggested Method(s)
What do people think/feel/experience? Interviews, surveys, focus groups
How does a system work in practice? Ethnography, case study, algorithmic audit
Does X cause Y? Experiment, quasi-experiment
How prevalent is a phenomenon? Survey, content analysis, statistical analysis
What does a document/policy/text say? Content analysis, legal analysis
How do multiple factors interact in a complex situation? Case study, mixed methods
What are the disparate impacts of a system? Algorithmic audit, statistical analysis

A practical recommendation: For most student research projects in data ethics, a case study combined with either content analysis or a small-scale survey will be the most feasible and productive approach. Large-scale experiments and algorithmic audits require resources and access that may be beyond the scope of a course project.


A.9 A Note on Interdisciplinarity

Data ethics draws on research traditions from computer science, law, philosophy, sociology, political science, economics, and science and technology studies (STS). Each tradition has its own standards for evidence, its own methodological norms, and its own criteria for what constitutes a "good" study. When reading across disciplines, be aware that:

  • A computer science paper may prioritize technical precision and reproducibility but underspecify social context.
  • A sociological study may provide rich contextual understanding but lack the statistical power to make generalizable claims.
  • A legal analysis may rigorously interpret existing law but underestimate the gap between law on the books and law in practice.
  • A philosophical argument may clarify values and principles but offer limited empirical evidence.

The most impactful work in data ethics -- the studies cited most frequently in this textbook -- typically bridges disciplines. Buolamwini and Gebru's Gender Shades study combined computer science benchmarking with sociological analysis of intersectional bias. Obermeyer et al.'s healthcare algorithm study combined statistical analysis with structural analysis of racial inequality in healthcare access. Aspire to this kind of integration in your own work.


Further Reading

  • Creswell, J.W. and Creswell, J.D. (2023). Research Design: Qualitative, Quantitative, and Mixed Methods Approaches. 6th ed. Sage.
  • Salganik, M.J. (2018). Bit by Bit: Social Research in the Digital Age. Princeton University Press. (Available free online.)
  • Metcalf, J. and Crawford, K. (2016). "Where are human subjects in Big Data research? The emerging ethics divide." Big Data & Society, 3(1).
  • Zook, M., Barocas, S., boyd, d., et al. (2017). "Ten simple rules for responsible big data research." PLoS Computational Biology, 13(3).