Skip to main content
M&E Studio
Home
Services
Tools
AI for M&E
Workflows
Plugins
Prompts
Start a Conversation
Library
Contribution AnalysisDevelopmental EvaluationImpact EvaluationLogframe / Logical FrameworkMost Significant ChangeOutcome HarvestingOutcome MappingParticipatory EvaluationProcess TracingQuasi-Experimental DesignRealist EvaluationResults FrameworkResults-Based ManagementTheory of ChangeUtilization-Focused Evaluation
M&E Studio

Decision-Grade M&E, Responsibly Built

About

  • About Us
  • Contact
  • LinkedIn

Services

  • Our Services
  • Tools

AI for M&E

  • Workflows
  • Plugins
  • Prompts
  • AI Course

M&E Library

  • Decision Guides
  • Indicators
  • Reference
  • Downloads

Legal

  • Terms
  • Privacy
  • Accessibility

© 2026 Logic Lab LLC. All rights reserved.

  1. M&E Library
  2. /
  3. Disaggregation
Core ConceptIndicators5 min read

Disaggregation

The breakdown of aggregate data by sub-group characteristics, such as sex, age, location, or vulnerability status, to reveal inequities and differences in programme reach and outcomes.

When to Use

Disaggregation should be built into every monitoring system where equity matters, which is virtually all development and humanitarian programmes. It becomes mandatory when:

  • Donors require sex-disaggregated data (USAID mandates sex disaggregation for all performance indicators; EU requires it under the Gender Action Plan)
  • The programme targets specific sub-groups (women, children, people with disabilities, ethnic minorities)
  • The theory of change explicitly aims to reduce inequities
  • Previous monitoring has shown that aggregate improvements masked disparities

Disaggregation is not just a technical practice, it is an equity commitment operationalised in data systems.

How It Works

Step 1: Identify required and meaningful disaggregation variables

Not every indicator needs every disaggregation variable. Decide which variables are mandatory (donor requirements), which are programme-relevant (age for a youth programme), and which are analytically meaningful (location for geographically targeted interventions). Standard disaggregation variables include:

  • Sex (mandatory for most donors): male, female, other/prefer not to say
  • Age group: child (0-17), youth (15-24), adult (25-59), older adult (60+)
  • Geographic location: district, urban/rural, programme zone
  • Wealth/vulnerability status: wealth quintile, food security status, displacement status
  • Disability status: WHO model disability survey categories

Step 2: Design data collection to capture sub-group data

Disaggregation requires that data collection instruments capture the sub-group variables for every respondent or unit of analysis. This means adding demographic questions to surveys, ensuring rosters capture sex and age, and training enumerators on consistent categorisation.

Step 3: Build disaggregation into the analysis plan

Specify which indicators will be analysed by which variables. Document this in the M&E plan and ensure the data management system can produce disaggregated tables.

Step 4: Report and act on disaggregated findings

Disaggregated data has no value if it stays in spreadsheets. Include disaggregated tables in programme reports and flag significant disparities. When disaggregation reveals that women, children, or a particular geographic group is underperforming relative to the aggregate, treat this as a programme management signal requiring response.

Key Components

  • Disaggregation plan: specifying which variables will be collected and which indicators will be disaggregated
  • Data collection instrument design: ensuring demographic variables are collected for all respondents
  • Reporting templates: standard tables showing aggregate and disaggregated results side by side
  • Minimum threshold guidance: the minimum sub-group sample size below which disaggregated results are not reported (typically n ≥ 30)
  • Equity analysis: comparison of outcomes across sub-groups to identify and address disparities

Best Practices

Disaggregate by sex as a minimum. Sex disaggregation is the most universal requirement and the most commonly missing. If you can only do one disaggregation, start here.

Collect data at the right level for disaggregation. Aggregate household data cannot be disaggregated to individual women's outcomes. Design data collection at the individual level if individual disaggregation is required.

Be consistent across time points. Disaggregation categories must be identical at baseline, midline, and endline to enable comparison.

Sample with sub-groups in mind. If a sub-group constitutes only 5% of the population but you need statistically meaningful results for them, you need to oversample that group. Random sampling of the full population will not produce adequate sub-group sample sizes.

Report disparity, not just disaggregated numbers. Saying "60% of women and 75% of men achieved the outcome" is more useful than listing both numbers in separate columns without comment.

Common Mistakes

Collecting disaggregation variables but not analysing them. Many programmes dutifully record sex and age in data collection but produce only aggregate numbers in analysis. Build disaggregated analysis into the reporting template so it cannot be skipped.

Insufficient sub-group sample sizes. When a sub-group has fewer than 30 respondents, statistical conclusions are unreliable. Plan sample sizes to enable meaningful sub-group analysis.

Too many disaggregation variables. Disaggregating every indicator by sex, age, location, and wealth simultaneously is analytically valuable but practically overwhelming. Prioritise which indicators need which disaggregation based on the programme's equity objectives.

Examples

Health programme, Sub-Saharan Africa. A PEPFAR-funded HIV prevention programme reported aggregate HIV testing rates in Year 1 that appeared strong (68% of target population tested). Disaggregation by age group revealed that testing rates among 15-24 year olds were only 41%, well below the programme target, while rates among adults 25-49 were 82%. This finding prompted targeted youth engagement activities that raised youth testing rates to 67% by Year 2.

Education programme, South Asia. A UNICEF-funded girls' education programme in Pakistan disaggregated school attendance data by wealth quintile in addition to sex. The data revealed that girls from the poorest quintile had attendance rates 30 percentage points lower than girls from the middle quintile, despite the programme providing stipends to all girls. Investigation revealed that stipend payment delays were disproportionately affecting the most remote villages, a logistical issue that was corrected in the subsequent term.

Related Topics

  • Gender-Responsive M&E, the broader framework for integrating gender equity into M&E systems
  • Indicator Selection, selecting indicators that can be meaningfully disaggregated
  • Target Setting, setting sub-group-specific targets to hold programmes accountable for equity
  • Baseline Design, designing baseline data collection to capture sub-group variables
  • SMART Indicators, indicators need to be measurable at sub-group level to be disaggregable

At a Glance

Breaks down monitoring and evaluation data by population sub-groups to identify who is and is not benefiting, and to hold programmes accountable for equitable outcomes.

Best For

  • Equity-focused programmes with gender, age, or inclusion mandates
  • Donor requirements for sex-disaggregated data (USAID, DFID, EU)
  • Identifying which sub-groups are underserved or experiencing worse outcomes
  • Informing targeting and resource allocation decisions

Complexity

Low to Medium

Timeframe

Designed during indicator development; data collected throughout programme

Linked Indicators

47 indicators across 5 donor frameworks

USAIDDFIDUNICEFUN WomenEU

Examples

  • Percentage of indicators reported with sex disaggregation
  • Outcome scores by age group, comparing youth versus adult beneficiaries
  • Service coverage rate by wealth quintile (poorest versus wealthiest)

Related Topics

Core Concept
SMART Indicators
A quality framework for designing indicators that are Specific, Measurable, Achievable, Relevant, and Time-bound, ensuring they provide reliable, actionable data for decision-making.
Core Concept
Indicator Selection & Development
The systematic process of choosing and refining performance indicators that are specific, measurable, achievable, relevant, and time-bound to track programme progress effectively.
Core Concept
Baseline Design
A structured approach to collecting initial condition data that directly informs project decisions, minimizes burden, and enables valid comparison with endline measurements.
Core Concept
Data Quality Assurance
A systematic process for verifying that collected data meets five quality dimensions, Validity, Integrity, Precision, Reliability, and Timeliness, ensuring data is fit for decision-making.
Core Concept
Gender-Responsive M&E
An approach to monitoring and evaluation that systematically examines how programmes affect women, men, girls, and boys differently, and ensures that M&E processes themselves do not reinforce gender inequalities.
Core Concept
Target Setting
The process of establishing specific, time-bound performance benchmarks against which programme progress and achievement will be measured.
Core Concept
Survey Design
The process of designing structured questionnaires and survey protocols to collect reliable, valid, and actionable data from a defined population.