Skip to main content
M&E Studio
AI for M&E
AI How-TosPromptsPlaybooksPlugins
Indicators
Workflows
M&E Resources
M&E MethodsReference Library
About
Services
FR — FrançaisES — Español
M&E Studio

AI for M&E, Built for Practitioners

AI for M&E

  • AI How-Tos
  • Prompts
  • Playbooks
  • Plugins
  • Indicators
  • Workflows

M&E Resources

  • M&E Methods
  • Reference Library
  • Decision Guides
  • Tools
  • Courses

Company

  • About
  • Services
  • Contact
  • LinkedIn

Legal

  • Terms
  • Privacy
  • Accessibility

© 2026 Logic Lab LLC. All rights reserved.

Library
  1. M&E Library
  2. /
  3. Disaggregation

Disaggregation

The breakdown of aggregate data by sub-group characteristics, such as sex, age, location, or vulnerability status, to reveal inequities and differences in program reach and outcomes.

When to Use

Disaggregation should be built into every monitoring system where equity matters - which is virtually all development and humanitarian programs. It becomes mandatory when:

  • Donors require sex-disaggregated data (USAID mandates sex disaggregation for all performance indicators; EU requires it under the Gender Action Plan)
  • The program targets specific sub-groups (women, children, people with disabilities, ethnic minorities)
  • The theory of change explicitly aims to reduce inequities
  • Previous monitoring has shown that aggregate improvements masked disparities

Disaggregation is not just a technical practice - it is an equity commitment operationalised in data systems.

How It Works

Step 1: Identify required and meaningful disaggregation variables

Not every indicator needs every disaggregation variable. Decide which variables are mandatory (donor requirements), which are program-relevant (age for a youth program), and which are analytically meaningful (location for geographically targeted interventions). Standard disaggregation variables include:

  • Sex (mandatory for most donors): male, female, other/prefer not to say
  • Age group: child (0-17), youth (15-24), adult (25-59), older adult (60+)
  • Geographic location: district, urban/rural, program zone
  • Wealth/vulnerability status: wealth quintile, food security status, displacement status
  • Disability status: WHO model disability survey categories

Step 2: Design data collection to capture sub-group data

Disaggregation requires that data collection instruments capture the sub-group variables for every respondent or unit of analysis. This means adding demographic questions to surveys, ensuring rosters capture sex and age, and training enumerators on consistent categorisation.

Step 3: Build disaggregation into the analysis plan

Specify which indicators will be analyzed by which variables. Document this in the M&E plan and ensure the data management system can produce disaggregated tables.

Step 4: Report and act on disaggregated findings

Disaggregated data has no value if it stays in spreadsheets. Include disaggregated tables in program reports and flag significant disparities. When disaggregation reveals that women, children, or a particular geographic group is underperforming relative to the aggregate, treat this as a program management signal requiring response.

Key Components

  • Disaggregation plan: specifying which variables will be collected and which indicators will be disaggregated
  • Data collection instrument design: ensuring demographic variables are collected for all respondents
  • Reporting templates: standard tables showing aggregate and disaggregated results side by side
  • Minimum threshold guidance: the minimum sub-group sample size below which disaggregated results are not reported (typically n ≥ 30)
  • Equity analysis: comparison of outcomes across sub-groups to identify and address disparities

Best Practices

Disaggregate by sex as a minimum. Sex disaggregation is the most universal requirement and the most commonly missing. If you can only do one disaggregation, start here.

Collect data at the right level for disaggregation. Aggregate household data cannot be disaggregated to individual women's outcomes. Design data collection at the individual level if individual disaggregation is required.

Be consistent across time points. Disaggregation categories must be identical at baseline, midline, and endline to enable comparison.

Sample with sub-groups in mind. If a sub-group constitutes only 5% of the population but you need statistically meaningful results for them, you need to oversample that group. Random sampling of the full population will not produce adequate sub-group sample sizes.

Report disparity, not just disaggregated numbers. Saying "60% of women and 75% of men achieved the outcome" is more useful than listing both numbers in separate columns without comment.

Common Mistakes

Collecting disaggregation variables but not analyzing them. Many programs dutifully record sex and age in data collection but produce only aggregate numbers in analysis. Build disaggregated analysis into the reporting template so it cannot be skipped.

Insufficient sub-group sample sizes. When a sub-group has fewer than 30 respondents, statistical conclusions are unreliable. Plan sample sizes to enable meaningful sub-group analysis.

Too many disaggregation variables. Disaggregating every indicator by sex, age, location, and wealth simultaneously is analytically valuable but practically overwhelming. Prioritize which indicators need which disaggregation based on the program's equity objectives.

Examples

Health program, Sub-Saharan Africa. A PEPFAR-funded HIV prevention program reported aggregate HIV testing rates in Year 1 that appeared strong (68% of target population tested). Disaggregation by age group revealed that testing rates among 15-24 year olds were only 41% - well below the program target - while rates among adults 25-49 were 82%. This finding prompted targeted youth engagement activities that raised youth testing rates to 67% by Year 2.

Education program, South Asia. A UNICEF-funded girls' education program in Pakistan disaggregated school attendance data by wealth quintile in addition to sex. The data revealed that girls from the poorest quintile had attendance rates 30 percentage points lower than girls from the middle quintile, despite the program providing stipends to all girls. Investigation revealed that stipend payment delays were disproportionately affecting the most remote villages - a logistical issue that was corrected in the subsequent term.

Proposal Context

Disaggregation commitments in proposals reflect both donor compliance (sex disaggregation is mandatory for USAID, FCDO, and EU; SDG reporting uses multiple disaggregation variables) and equity-focused programming. Common proposal pitfalls: (a) promising disaggregation without sample sizes to support it (n=300 disaggregates by sex acceptably but not by age-by-sex with useful precision), (b) performative disaggregation (just sex, or just age) without meaningful equity analysis (disparities identified but not acted on), (c) inconsistent disaggregation across baseline, midline, and endline (categories change, comparability breaks), (d) no analysis plan for disaggregated findings (disaggregated data collected and stored but not reported), (e) disaggregation categories that do not match donor requirements (donor requires standard age bands; proposal uses different ones). A strong proposal commits to specific disaggregation variables, specifies sample sizes that support them, and plans equity analysis. Pair with smart-indicators and sampling-methods.

Related Topics

  • Gender-Responsive M&E: the broader framework for integrating gender equity into M&E systems
  • Indicator Selection: selecting indicators that can be meaningfully disaggregated
  • Target Setting: setting sub-group-specific targets to hold programs accountable for equity
  • Baseline Design: designing baseline data collection to capture sub-group variables
  • SMART Indicators: indicators need to be measurable at sub-group level to be disaggregable

At a Glance

Breaks down monitoring and evaluation data by population sub-groups to identify who is and is not benefiting, and to hold programs accountable for equitable outcomes.

Best For

  • Equity-focused programs with gender, age, or inclusion mandates
  • Donor requirements for sex-disaggregated data (USAID, DFID, EU)
  • Identifying which sub-groups are underserved or experiencing worse outcomes
  • Informing targeting and resource allocation decisions

Linked Indicators

47 indicators across 5 donor frameworks

USAIDDFIDUNICEFUN WomenEU

Examples

  • Percentage of indicators reported with sex disaggregation
  • Outcome scores by age group, comparing youth versus adult beneficiaries
  • Service coverage rate by wealth quintile (poorest versus wealthiest)

Related Topics

Overview
SMART Indicators
A quality framework for designing indicators that are Specific, Measurable, Achievable, Relevant, and Time-bound, ensuring they provide reliable, actionable data for decision-making.
Overview
Indicator Selection & Development
The systematic process of choosing and refining performance indicators that are specific, measurable, achievable, relevant, and time-bound to track program progress effectively.
Overview
Baseline Design
A structured approach to collecting initial condition data that directly informs project decisions, minimizes burden, and enables valid comparison with endline measurements.
Overview
Data Quality Assurance
A systematic process for verifying that collected data meets five quality dimensions, Validity, Integrity, Precision, Reliability, and Timeliness, ensuring data is fit for decision-making.
Overview
Gender-Responsive M&E
An approach to monitoring and evaluation that systematically examines how programs affect women, men, girls, and boys differently, and ensures that M&E processes themselves do not reinforce gender inequalities.
Overview
Target Setting
The process of establishing specific, time-bound performance benchmarks against which program progress and achievement will be measured.
Overview
Survey Design
The process of designing structured questionnaires and survey protocols to collect reliable, valid, and actionable data from a defined population.

Decision Guides

How to Choose Sample Size for M&E
A practical guide to sample size for program evaluations, with rules of thumb, worked examples, and budget-statistics tradeoffs.
PreviousCustom vs Standard IndicatorsNextEndline