Accredited official statistics

Background information: Fraud and error in the benefit system statistics, 2023 to 2024 estimates

Published 16 May 2024

Applies to England, Scotland and Wales

Purpose of the statistics

Context and purpose of the statistics

This document supports our main publication which contains estimates of the level of fraud and error in the benefit system in Financial Year Ending (FYE) 2024.

We measure fraud and error so we can understand the levels, trends, and reasons behind it. This understanding supports decision making on what actions DWP can take to reduce the level of fraud and error in the benefit system. The National Audit Office takes account of the amount of fraud and error when they audit DWP’s accounts each year.

Within DWP these statistics are used to evaluate, develop, and support fraud and error policy, strategy and operational decisions, initiatives, options, and business plans through understanding the causes of fraud and error.

The fraud and error statistics published in May each year feed into the DWP accounts. The FYE 2024 estimates published in May 2024 feed into the FYE 2024 DWP annual report and accounts.

The statistics are also used within the annual HM Revenue and Customs National Insurance Fund accounts. These are available in the National Insurance Fund Accounts section of the HMRC reports page.

The fraud and error estimates are also used to answer Parliamentary Questions and Freedom of Information requests, and to inform DWP Press Office statements on fraud and error.

Limitations of the statistics

The estimates do not include reviews of every benefit each year. This year, Employment and Support Allowance was removed from the measured list of benefits and Disability Living Allowance was measured for the first time since FYE 2005. Also, this year the Pension Age Housing Benefit client groups were measured rather than the Non-passported Working Age Housing Benefit client group that was measured in FYE 2023.

This document includes further information on limitations – for example, on benefits reviewed and changes this year (section 1 and section 2, omissions to the estimates (section 3), and our sampling approach (section 4).

Longer time series comparisons may not be possible for some levels of reporting due to methodology changes. Our main publication and reference tables indicate when comparisons should not be made.

We are unable to provide sub-national estimates of fraud and error as we are unable to break the statistics down to this level.

Comparisons between the statistics

These statistics relate to the levels of fraud and error in the benefit system in Great Britain.

Social Security Scotland report the levels of fraud and error for benefit expenditure devolved to the Scottish Government within their annual report and accounts.

Northern Ireland fraud and error statistics are comparable to the Great Britain statistics within this report as their approach to collecting the measurement survey data, and calculating the estimates and confidence intervals, is very similar. Northern Ireland fraud and error in the benefit system high level statistics are published within the Department for Communities annual reports.

HM Revenue and Customs produce statistics on error and fraud in Tax Credits.

When comparing different time periods within our publication, we recommend comparing percentage rates of fraud and error rather than monetary amounts. This is because the amount of fraud and error in pounds could go up, even if the percentage rate of fraud and error stays the same or goes down, if the amount of benefit we pay out in total goes up compared to the previous year.

Source of the statistics

We take a sample of benefit claims from our administrative systems. DWP’s Performance Measurement (PM) team contact the benefit claimants to arrange a review. The outcomes of these reviews are recorded on a bespoke internal database called FREDA. We use data from here to produce our estimates.

We also use other data to inform our estimates – for example:

  • benefit expenditure data (aligning with the Spring Budget published forecasts).
  • benefit recovery data (DWP benefits and Housing Benefit) to allow us to calculate estimates of net loss.
  • other DWP data sources and models to improve the robustness of, or categorisations within, our estimates – for example, to allow us to see if claimants who leave benefit as a consequence of the fraud and error review process then return to benefit shortly afterwards, and to understand the knock-on effect of fraud and error on disability benefits on other benefits.

Further information on the data we use to produce our estimates is contained within section 4, section 5 and section 6 of this report.

Definitions and terminology within the statistics

The main publication presents estimates of Fraud, Claimant Error, and Official Error. The definitions for these are as follows:

  • Fraud: This includes all cases where the following three conditions apply:
    • the conditions for receipt of benefit, or the rate of benefit in payment, are not being met
    • the claimant can reasonably be expected to be aware of the effect on entitlement
    • benefit stops or reduces as a result of the review
  • Claimant Error: The claimant has provided inaccurate or incomplete information, failed to report a change in their circumstances, or failed to provide requested evidence, but there is no fraudulent intent on the claimant’s part.
  • Official Error: The benefit has been paid incorrectly due to a failure to act, a delay or a mistaken assessment by DWP, a local authority or HMRC, to which no one outside of that department has materially contributed

We report overpayments (where we have paid people too much money), and underpayments (where we have not paid people enough money).

We present these in percentage terms (of expenditure on a benefit) and in monetary terms, in millions of pounds.

We also report a measure on the percentage of cases with Fraud or an overpayment Error calculated as follows:

Proportion of claims with an overpayment or underpayment (reference tables 12 and 13):

Proportion of claims with Fraud or an overpayment error = (number of claims in the sample with at least one Fraud or at least one overpayment error) / (number of claims in the sample)

Proportion of claims with an underpayment error = (number of claims in the sample with at least one underpayment error) / (number of claims in the sample)

Since the same claim can be included in both the proportion of claims with an overpayment error or Fraud and the proportion of claims with an underpayment error, these figures cannot be summed together to obtain the total proportion of claims paid incorrectly.

Proportion of claims paid the incorrect amount (reference table 11):

Proportion of claims overpaid = (number of claims in the sample ultimately overpaid) / (number of claims in the sample)

Proportion of claims underpaid = (number of claims in the sample ultimately underpaid) / (number of claims in the sample)

These figures can be summed together to obtain the total proportion of claims paid incorrectly. Further information about the types of errors we report on, abbreviations commonly used and statistical methodology can be found in the appendices at the end of this document.

Revisions to the statistics

Revisions to our statistics may happen for a number of reasons. When we make methodology changes that impact our estimates, we may revise the estimates for the previous year to allow meaningful comparisons between the two. Where we introduce major changes, we may denote a break in our time series and recommend that comparisons are not made back beyond a certain point.

In our FYE 2024 publication we have revised:

  • the monetary value and rates of overpayments across the benefits for FYE 2023
  • the proportion of cases overpaid and incorrect across the benefits for FYE 2023
  • the proportion of Universal Credit (UC) cases with at least one Fraud overpayment in FYE 2023

For more information and the reason behind the revisions please see section 2.

The National Statistics Code of Practice allows for revisions of figures under controlled circumstances: “Statistics are by their nature subject to error and uncertainty. Initial estimates are often systematically amended to reflect more complete information. Improvements in methodologies and systems can help to make revised series more accurate and more useful.”

Unplanned revisions of figures in reports in this series might be necessary from time to time. Under this Code of Practice, the Department has a responsibility to ensure that any revisions to existing statistics are robust and are freely available, with the same level of supporting information as new statistics.

Status of the statistics

National statistics

National Statistics are accredited official statistics. National Statistics status means that these statistics meet the highest standards of trustworthiness, quality and public value, and there is a responsibility to maintain compliance with these standards.

These official statistics were independently reviewed by the Office for Statistics Regulation in December 2017. They comply with the standards of trustworthiness, quality and value in the Code of Practice for Statistics and should be labelled ‘accredited official statistics’. Our statistical practice is regulated by the Office for Statistics Regulation (OSR).

OSR sets the standards of trustworthiness, quality and value in the Code of Practice for Statistics that all producers of official statistics should adhere to.

You are welcome to contact us directly at enquiries.fema@dwp.gov.uk with any comments about how we meet these standards.

Alternatively, you can contact OSR by emailing regulation@statistics.gov.uk or via the OSR website.

Read further information about National Statistics on the UK Statistics Authority website.

Quality Statement

Quality in statistics is a measure of their ‘fitness for purpose’. The European Statistics System Dimensions of Quality provide a framework in which statisticians can assess the quality of their statistical outputs. These dimensions of quality are relevance, accuracy and reliability, timeliness, accessibility and clarity, and comparability and coherence.

Section 6 gives information on the application of these quality dimensions to our fraud and error statistics.

Feedback

We welcome any feedback on our publication. You can contact us at: enquiries.fema@dwp.gov.uk

Lead Statistician: Adam Pearce

DWP Press Office: 020 3267 5144

Report Benefit Fraud: 0800 854 4400

Releases of Fraud and error in the benefit system statistics.

1. Introduction to our measurement system

The main statistical release and reference tables and charts provide estimates of fraud and error for benefit expenditure administered by the Department for Work and Pensions (DWP). This includes a range of benefits for which we derive estimates using different methods, as detailed below. For further details on which benefits are included in the total fraud and error estimates please see Appendix 2. More information can be found on the GOV.UK website about the benefit system and how DWP benefits are administered.

The fraud and error estimates provide estimates for the amount overpaid or underpaid in total and by benefit, broken down into the types of Fraud, Claimant Error and Official Error for benefits reviewed this year.

Estimates of fraud and error for each benefit have been derived using three different methods, depending on the frequency of their review (see section 5 for details):

Benefits reviewed this year

Fraud, Claimant Error and Official Error (see definitions above) have been measured for FYE 2024 for Universal Credit (UC), Housing Benefit (HB), Disability Living Allowance (DLA), Pension Credit (PC), State Pension (SP) and Personal Independence Payment (PIP).

Expenditure on measured benefits accounted for 83% of all benefit expenditure in FYE 2024.

Estimates are produced by statistical analysis of data collected through annual survey exercises, in which independent specially trained staff from the Department’s Performance Measurement (PM) team review a randomly selected sample of cases for benefits reviewed this year. See section 4 for more information on the sampling process.

The review process involves the following activity:

  • previewing the case by collating information from a variety of DWP or Local Authority (LA) systems to develop an initial picture and to identify any discrepancies between information from different sources

  • interviewing the claimant (or a nominated individual where the claimant lacks capacity) using a structured and detailed set of questions about the basis of their claim. The interview is completed as a telephone review in the majority of cases. However, where this is not appropriate, there is usually also the option for a completed review form to be returned by post

  • the interview aims to identify any discrepancies between the claimant’s current circumstances and the circumstances upon which their benefit claim was based

If a suspicion of Fraud is identified, an investigation is undertaken by a trained Fraud Investigator with the aim of resolving the suspicion.

Benefits were measured with different sample periods, although all were contained within the period November 2022 to October 2023, except for PIP and DLA, which ran from November 2022 to August 2023 and Feb 2023 to September 2023 respectively. For more information on the sample period for individual benefits please see Annex 1 of the statistical report. The following number of benefit claims were sampled and reviewed by the PM team:

Benefit Sample size Percentage of claimant population reviewed
Universal Credit 3,989 0.09%
State Pension 1,557 0.01%
Housing Benefit 3,001 0.13%
Pension Credit 1,989 0.15%
Daily Living Allowance 1,385 0.11%
Personal Independence Payment 1,395 0.06%
Total 13,313 0.06%

Overall, approximately 0.06% of all benefit claims in payment were reviewed by the PM team.

Read information about the Performance Measurement Team.

Benefits reviewed previously

Since 1995, the Department has carried out reviews for various benefits to estimate the level of fraud and error in a particular financial year following the same process outlined above. In FYE 2024 around 11% of total expenditure related to benefits reviewed in previous years. Please see Appendix 2 for details of benefits reviewed previously.

Benefits never reviewed

The remaining benefits, which account for around 6% of total benefit expenditure, have never been subject to a specific review. These benefits tend to have relatively low expenditure which means it is not cost effective to undertake a review. For these benefits the estimates are based on assumptions about the likely level of fraud and error (for more information please see section 5).

2. Changes to the statistics this year

This section provides detail of changes for the FYE 2024 publication. Any historical changes can be found in Appendix 5.

Changes

Changes to benefits reviewed

Each year we use decision making methodology called multiple-criteria decision analysis (MCDA) to help evaluate which benefits will be reviewed.

For FYE 2024 we have measured DLA for the first time since FYE 2005 and stopped the measurement of ESA.

Unfulfilled Eligibility

From this year onward, estimates that were previously published as Claimant Error underpayments within the fraud and error statistics have been reclassified and are published separately as Estimates of Unfulfilled Eligibility. This follows a planned review of the fraud and error statistics to align the statistics more closely with benefit legislation. The review determined that the estimates previously published as Claimant Error underpayments should not be defined as underpayments. In benefit legislation, claimants are not eligible for increases in their benefit until they accurately report their circumstances to the department.

Unfulfilled eligibility refers to claimants who are already in receipt of a certain benefit but may not be getting the full award they could be eligible for on this benefit. This is not the same as take-up of benefits, which is where people could have claimed certain benefits based on their current circumstances but have not done so. The department also publishes statistics about take-up of benefits. For more information, please go to our publication regarding Estimates of Unfulfilled Eligibility.

Extra check on Failure to provide evidence cases

This relates to cases where the claimant has participated in the benefit review but has then failed to send in the requested evidence and subsequently had their claim to benefit terminated for failing to comply.

There is already an adjustment done on these cases where we look four months after their claim was terminated to see if we can find any evidence that would allow us to reclassify the fraud to a known error reason. This year we expanded this adjustment to see if we can find any evidence that would allow us to remove the fraud entirely on some of these cases (for example if we can see evidence of earnings coming through RTI the month after the benefit was terminated, which would have meant their Universal Credit entitlement went to zero). Applying these extra checks to the FYE 2023 statistics removed just over £100m of the “Failure to Provide Evidence/Engage” Fraud on Universal Credit. All other benefits were unaffected.

Change to the recording of Deemed Errors

In previous years, deemed errors were recorded, which are official errors where evidence that was available to the department when the award of benefit was made has been misplaced or is not available. When this happened, it means that the official error check could not be completed. Any case that had a deemed error raised against it was excluded from the calculation of the official error rate (but may still be counted in the calculation of Claimant Error and Fraud).

This year rather than record a deemed error, an estimated outcome is recorded instead, and the case is retained in the calculation of Official Error. We estimate that had this change been carried out last year it would not have changed the reported figures.

3. Interpretation of the results

Care is required when interpreting the results presented in the main report:

  • the estimates are based on a random sample of the total benefit caseload and are therefore subject to statistical uncertainties. This uncertainty is quantified by the estimation of 95% confidence intervals surrounding the estimate. These 95% confidence intervals show the range within which we would expect the true value of fraud and error to lie
  • when comparing two estimates, users should take into account the confidence intervals surrounding each of the estimates. The calculation to determine whether the results are significantly different from each other is complicated and takes into account the width of the confidence intervals. We perform this robust calculation in our methodology and state in the report whether any differences between reporting years are significant or not
  • unless specifically stated within the commentary in the publication or in the reference tables, none of the changes for benefits reviewed this year are statistically significant at a 95% level of confidence when compared to the previous measurement

As well as sampling variation, there are many factors that may also impact on the reported levels of fraud and error and the time series presented:

  • these estimates are subject to statistical sampling uncertainties. All estimates are based on reviews of random samples drawn from the benefit caseloads. In any survey sampling exercise, the estimates derived from the sample may differ from what we would see if we examined the whole caseload. Further uncertainties occur due to the assumptions that have had to be made to account for incomplete or imperfect data or using older measurements
  • the sample year and the financial year do not align. This means that a proportion of expenditure for benefits reviewed this year cannot be captured by the sampling process. This is mainly because of the delay between sample selection and the interview of the claimant, and also the time taken to process new benefit claims, which excludes the newest cases from the review. The estimates in the reference tables in this release have been extrapolated to account for the newest benefit claims which are missed in the benefit reviews and cover all expenditure
  • the estimates do not encompass all fraud and error. This is because Fraud is, by its nature, a covert activity, and some suspicions of Fraud on the sample cases cannot be proven. For example, cash in hand earnings are harder to detect than those that get paid via PAYE. Complex official error can also be difficult to identify. More information on omissions can be found later in this section
  • some incorrect payments may be unavoidable. The measurement methodology will treat a case as incorrect, even where the claimant has promptly reported a change and there is only a short processing delay

Omissions from the estimates

The fraud and error estimates do not capture every possible element of fraud and error. Some cases are not reviewed due to the constraints of our sampling or reviewing regimes (or it is impractical to do so from a cost or resource perspective), some cases are out of scope of our measurement process, and some elements are very difficult for us to detect during our benefit reviews. The time period that our reviews relate to means that any operational or policy changes in the last five months of the financial year are usually not covered by our measurements.

For most omissions from our estimates, we make adjustments or apply assumptions to those cases. For some omissions we assume that the levels of fraud and error for those cases are the same as for the cases that we do review, and for other omissions we apply specific assumptions where we expect the levels of fraud and error to be different.

This section details the omissions from the estimates as far as possible. The examples that follow are not an exhaustive list but are an attempt at providing further details on known omissions in the estimates.

There are a number of groups of cases that we are unable to review or which we do not review. Some of the main examples of these are as follows:

New and short-term cases

We are unable to review short duration cases (of just a few weeks in duration) due to the time lags involved in accessing data on the benefit caseloads, drawing the samples and preparing these for reviewing. For these cases, we assume the rates of fraud and error are the same as in the rest of the benefit caseloads. We do, however, also make an adjustment using “new cases factors” to try to ensure that the results are representative across the entire distribution of lengths of benefit claims (see section 5 for further details).

It can take time for new cases to be available for sampling, meaning they are potentially under-represented in the sample. Analysis was undertaken to quantify the impact of these potential exclusions.

New cases make up a small proportion of cases for most benefits. The table below shows the yearly average percentage that are less than three months old at a given time.

Benefit Average number of cases less than 3 months old Source
Pension Credit 2.5% Official quarterly data for August 2022 to May 2023
Disability Living Allowance 1.7% Official quarterly data for August 2022 to May 2023
Personal Independence Payment 4.1% Official monthly data for November 2022 to October 2023
Housing Benefit 3.9% Official monthly new case data September 2022 to August 2023
Universal Credit 7.8% Official monthly data for December 2022 to November 2023
State Pension 0.7% Estimated using quarterly pension data for May 2022 and ONS population data

To investigate the impact of this exclusion, a simulation of the sampling process was performed and repeated multiple times with these cases included. Sensitivity analysis was then carried out across all benefits to estimate the impact of excluding new cases if the error rate were doubled, halved, or remained the same.

The age of a case when it becomes available for sampling differs by benefit. Based on these timescales, analyses were undertaken using an exclusion period of either 6 weeks or 3 months.

Overall, this analysis showed that, because short term cases make up such a small percentage of total cases at any given time and many are available to be sampled later in the period, the impact on final published figures for all benefits is negligible.

The impact of excluding new cases is no more than 0.2 percentage points difference in the estimated error rate for either underpayment or overpayment across all benefits except UC. For UC overpayments, the scenario of doubling the error rate increases the total overpayment rate by 0.4 percentage points and halving the error rate decreases it by 0.2 percentage points. The higher numbers reflect the fact that UC has a higher proportion of new cases and a higher rate of error than other benefits.

It would be expected that the rate of fraud and error in new cases would typically be lower than the full population of claimants since they have recently been assessed. The outcomes of this analysis fall within the estimated confidence intervals and there is little impact on the published statistics, therefore no adjustment is required.

Unclaimed Benefits

We are only able to sample claimants who are in receipt of a benefit payment. Eligible claimants that have not made a benefit claim are not included in these figures. Find statistics on benefit take-up.

Disallowed claims

Claims which do not receive an award are not included in our sample. These claims may have been disallowed in error, resulting in a possible underpayment.

Using data on the number of disallowed claims and appeal success rates, sensitivity analysis was carried out to investigate the impact of various assumed error rates for disallowed cases.

A summary across benefits for various error rate scenarios is shown below. The worst-case scenario is not plausible and is included to illustrate that even assuming an unrealistic level of error, the adjusted estimate would still fall within or close to the published confidence intervals.

Results show that there is likely to be little to no impact on our estimate of underpayments as a result of not sampling disallowed claims.

Disability Living Allowance

Level of underpayment in excluded group Estimated change in underpayment Impact on DLA underpayment estimate Impact on global underpayment estimate
Extreme worst case: appeal success rate applied to all disallowed claims + £44.4m + 0.7 p.p + 0.0 p.p
Disallowed cases have the same error rate as measured cases + £2.6m + 0.0 p.p + 0.0 p.p

Personal Independence Payment

Level of underpayment in excluded group Estimated change in underpayment Impact on PIP underpayment estimate Impact on global underpayment estimate
Extreme worst case: All disallowed cases are appealed with the same success rate + £238.7m + 1.1 p.p + 0.1 p.p
Disallowed cases have the same error rate as measured cases + £11.6m + 0.1 p.p + 0.0 p.p

Note: PIP worst case analysis is slightly different due to its high appeal rate.

Pension Credit

The data on disallowed cases shows a steep increase in the proportion of cases declined with no award which happened in FYE 2021 and is continuing into the current financial year. Rejection rates in FYE 2021 and FYE 2022 have been affected by the financial strain people experienced due to the pandemic, causing more ineligible people to try and claim the benefit. Furthermore, following the recent government campaign published in April 2022, urging eligible pensioners to apply for Pension Credit, there has been an even larger increase in the volume of claims received by Department for Work and Pensions. In the meantime, many financial journalists started encouraging people to apply without checking the eligibility criteria due to the complex nature of the benefit. This meant that many people who were not eligible for Pension Credit tried to claim anyway, naturally increasing the rejection rate throughout FYE 2023 and FYE 2024. Therefore, we were not able to use this year’s and last year’s rejection rate in our sensitivity analysis and the last accurate rejection rate for our sensitivity analysis has been obtained from FYE 2020. If we used the most recent rejection rate, we would be overestimating the impact due to a larger number of ineligible claimants being rejected with no award.

Level of underpayment in excluded group Estimated change in underpayment Impact on PC underpayment estimate Impact on global underpayment estimate
Extreme worst case: appeal success rate applied to all disallowed claims + £21.7m + 0.4 p.p + 0.0 p.p
Disallowed cases have the same error rate as measured cases + £4.9 m + 0.1 p.p + 0.0 p.p
Twice as many cases are eligible to be awarded on reconsideration as actually are + £1.1m +0.0 p.p + 0.0 p.p

Universal Credit

Appeals information was not available for UC therefore the sensitivity analysis could not be carried out fully as in the above tables.

Using numbers of disallowed cases, analysis shows that even if the error rate in disallowed cases was double that found in the sample, the total difference in the UC rate would be around £7.7 million, or 0.0% of expenditure, and therefore well within the confidence interval – no adjustment is appropriate.

State Pension

Recent data was not available on the number of disallowed State Pension claims, however, given the data available it is expected to be very low. For the most recent year for which data on appeal rates was available (November 2022 to October 2023),applying sensitivity analysis based on a series of reasonable worst-case assumption showed that even if the error rate in disallowed cases was double that found in the sample, the total difference in the State Pension rate would be £0.2 million, which is negligible. Looking at the worst-case scenario, where all disallowed claims (even the ones that were not challenged) are reconsidered with the same error rate that was found in the sample, the total difference in the State Pension rate would be £4 million, or 0.0% of the expenditure. Therefore, no adjustment is required.

Housing Benefit

Applications for Housing Benefit are processed by local authorities, not centrally by DWP as with the other benefits reviewed, and it was not possible to obtain numbers of disallowed cases.

Since analyses of other benefits has shown negligible impact in excluding disallowed claims, the assumption is made that the impact on Housing Benefit is also negligible.

In conclusion, no adjustments are required to the estimates to account for the exclusion of disallowed cases from the sample.

Nil payment claims

A case is considered to be nil-payment if there is a claim in place but the total award being paid is zero. These cases are not included when the sample is selected. Some benefits do not allow a nil-award, meaning all active claimants are receiving some payment.

Nil-award allowed No nil-award allowed
Pension Credit Housing Benefit
Universal Credit State Pension
- Personal Independence Payment

Note: For a very small proportion of the PIP caseload (0.2%), the combination of award rates (daily living and mobility) is reported as nil-nil. Investigations suggest that award rates may be temporarily shown as nil for a short period whilst a claim review is in process, after which the new award rate is set. These cases will be monitored.

Nil-payment claims are a potential source of underpayment that is not included in the sample.

Pension Credit

Only a very small percentage of Pension Credit claimants are in nil-payment at any given time. For the most recent year of data available this number was always below one fifth of one percent. Simulating the sampling process shows that over 99.8% of cases sampled would be the same even if the nil-payment cases were included in the group available for sampling. The potential impact of a different error rate in this group thus rounds to zero, even in a worst-case scenario of doubling the error rate in the excluded cases.

Universal Credit

For the year up until August 2023 the most recent year for which data was available for analysis, the number of nil-payment claims ranged from 9.5% to 12.3%, with 11.0% being the average. Simulating the sampling process shows around 11.0% of claims that would otherwise be sampled are missed due to this. The potential impact of a different error rate in this group is a decrease of 0.1 percentage points if the error rate is half the measured rate, and an increase of 0.2 percentage points if the error rate in this group is double the measured rate.

Exclusions specific to PIP

The monthly samples are taken from live PIP claims in advance of the scheduling of the benefit reviews. Any benefit record relating to a claimant who meets specific exclusion criteria (for example, terminally ill, reviewed in the last three months) will not be reviewed. We assume the rates of fraud and error for these cases are the same as the rest of the PIP caseload. The potential impact of each excluded group are summarised below.

Terminally Ill cases

Terminally ill claimants make up a very small percentage of PIP claimants, typically around 1%. Sensitivity analysis was carried out to test the impact of their exclusion if the fraud and error rate on these cases were to differ by as much as double that of the sampled population.

Estimated impact of excluding terminally ill cases on the PIP overpayment rate

Level of overpayment in excluded group Estimated change in overpayments Impact on PIP overpayment estimate Impact on global overpayment estimate
Double the published rate + £0.8m 0.0 p.p 0.0 p.p
Half the published rate - £0.4m 0.0 p.p 0.0 p.p

Estimated impact of excluding terminally ill cases on the PIP underpayment rate

Level of underpayment in excluded group Estimated change in underpayments Impact on PIP underpayment estimate Impact on global underpayment estimate
Double the published rate £0.0m 0.0 p.p 0.0 p.p
Half the published rate £0.0m 0.0 p.p 0.0 p.p

Note that the PIP underpayment analysis applies to cases that are not on the highest award (enhanced mobility with enhanced daily living) as the estimates are already adjusted to account for this exclusion. See the Benefit-specific adjustments section for more details.

The above analysis demonstrates that any additional fraud or error that may exist in the excluded PIP terminally ill cases is extremely small. The total estimated levels of fraud and error would have a negligible difference for the above scenarios if terminally ill cases were included in the sample. Therefore, no adjustment has been made to the statistics because of this exclusion.

Scheduled reviews

PIP awards are reviewed regularly. The period between reviews is set on an individual basis and ranges from 9 months to 10 years, with the majority of claimants having a short-term award of 0-2 years. If a claim has had a planned award review in the last 92 days, has a review ongoing or a review due in the next six weeks then it is not eligible to be sampled.

A simulation of the sampling process was performed and repeated multiple times to investigate the impact of this exclusion.

Cases with an upcoming review would be expected to have a higher propensity for Fraud or Error due to the length of time since their last review. By the same reasoning cases that have had a review completed recently would be expected to have a lower propensity for Fraud or Error.

In our statistics an assumption is made that the excluded cases are similar to those sampled and so no adjustments are made. We investigated the impact of alternative assumptions on our estimates and results are shown in the table below.

In these scenarios it was assumed that an award that was increased after a planned review would have an underpayment had we reviewed it, and cases that were decreased or reviewed before being disallowed would have an overpayment. For overpayments, these proportions were taken from PIP data published via StatXplore and fed into the results below.

Estimated impact of excluding cases with a review that is due, ongoing, or recently completed using review outcomes to estimate fraud and error rates

For PIP overpayments, the vast majority of error was Claimant Error. Therefore, for this analysis we assumed zero error on the recently reviewed claims that were excluded from the sample, given that a recent review would resolve potential Claimant Error. A worst-case scenario of half the sampled error existing in the excluded recently reviewed cases was also estimated.

Therefore, for overpayments two assumptions were tested for recently reviewed cases: that they have no errors, and that they have half the error rate as our sample. The table below shows these assumptions and their estimated difference to the published amounts and rates.

Overpayments

Estimated difference from published Impact on PIP estimate Impact on global estimate
Assumption of zero error on excluded recently reviewed cases £7.0m 0.3 p.p 0.0 p.p
Assumption of half the sampled error on excluded recently reviewed cases £14.7m 0.3 p.p 0.0 p.p

As this publication focuses on Official Error underpayments only, a different method of analysis on excluded cases was applied for underpayments. We were unable to split the available review outcome data down into type of error, therefore applied a simplified calculation based on the sampled Official Error estimates for underpayments.

For underpayments, we assumed that Official Error is likely to occur in the excluded recently reviewed cases in the same way as for our sampled cases. Therefore, we have estimated the impact of the excluded cases having half the error as the sample, and as a worst-case scenario having double the error as the sampled cases.

Therefore, for underpayments two assumptions were tested for recently reviewed cases: that they have half the sampled error, and that they have double the error as our sample. The table below shows these assumptions and their estimated difference to the published amounts and rates.

Underpayments

Estimated difference from published Impact on PIP estimate Impact on global estimate
Assumption of half the sampled error on excluded recently reviewed cases -£8.0m 0.0 p.p 0.0 p.p
Assumption of double the error on excluded recently reviewed cases £16.2m 0.1 p.p 0.0 p.p

These differences all fall within the published confidence intervals for the rates of overpayment and underpayment for PIP in FYE 2024, and the conclusion therefore is no adjustment is needed.

Exclusions specific to DLA

The monthly samples are taken from live DLA claims in advance of the scheduling of the benefit reviews. Any benefit record relating to a claimant who meets specific exclusion criteria (e.g., terminally ill, reviewed in the last three months) will not be reviewed. We assume the rates of fraud and error for these cases are the same as the rest of the DLA caseload. The potential impacts of each excluded group are summarised below.

Working age cases

DLA is being replaced by other benefits. New claims to DLA can only be made for children.

Adults of working age and pensioners aged 65 or over prior to 8 April 2013 (DLA 65+), already in receipt of DLA, will continue to get DLA as long as they remain eligible for it.

We have reviewed claims for children and DLA 65+ only. Working age cases are not being reviewed due to being unable to measure a change in rate – DLA working age cases with a change in circumstances have their DLA payments ended and must claim PIP instead. We exclude current pensioners who were under 65 on 8th April 2013 who would also have their DLA payment ended if a change of circumstances was found and they would have to claim PIP. Working age cases account for about 15% of all DLA cases. We considered what proxy to use for the unmeasured working age element of DLA. Analysis of rate combinations of each client group suggests that a combination of Children and 65+ rates, weighted by population factor, best match the working age claimants’ rate. However, lack of review for those cases prevents a more accurate assessment.

As DLA working age claims with a change in circumstances cannot be measured, we have excluded them from our sensitivity analyses and adjustments.

Terminally ill cases

Terminally ill claimants make up a very small percentage of DLA claimants, typically less than 1%.

Sensitivity analysis was carried out to test the impact of their exclusion if the fraud and error rate on these cases were to differ by as much as double that of the sampled population.

Estimated impact of excluding terminally ill cases on the DLA overpayment rate

Level of overpayment in excluded group Estimated change in overpayments Impact on DLA overpayment estimate Impact on global overpayment estimate
Double the published rate £0.1m 0.0 p.p 0.0 p.p
Half the published rate £0.0m 0.0 p.p 0.0 p.p

Estimated impact of excluding terminally ill cases on the DLA underpayment rate

Level of underpayment in excluded group Estimated change in underpayments Impact on DLA underpayment estimate Impact on global underpayment estimate
Double the published rate £0.0m 0.0 p.p 0.0 p.p
Half the published rate £0.0m 0.0 p.p 0.0 p.p

Note that DLA claims on the highest rate award (higher rate mobility and high care) have been excluded from this analysis as they can’t be underpaid.

The above analysis demonstrates that any additional fraud or error that may exist in the excluded DLA terminally ill cases is extremely small. The total estimated levels of fraud and error would have a negligible difference for the above scenarios if terminally ill cases were included in the sample. Therefore, no adjustment has been made to the statistics because of this exclusion.

Scheduled reviews

DLA claims for children are reviewed on renewal, however DLA 65+ awards are not routinely reviewed. Around 18% of the sample was abandoned due to cases having a scheduled review or a recent review, and they were therefore excluded from the sample estimates.

In our statistics an assumption is made that the excluded cases are similar to those sampled and so no adjustments are made. We investigated the impact of alternative assumptions on our estimates and results are shown in the table below.

Estimated impact of excluding cases with a review that is due, ongoing, or recently completed using review outcomes to estimate fraud and error rates

Overpayments

For DLA overpayments two assumptions were tested for cases excluded as recently reviewed or having a scheduled review. We have estimated the impact of the excluded cases having half the error as the sample, and as a worst-case scenario having double the error as the sampled cases. The table below shows these assumptions and their estimated difference to the published amounts and rates.

Estimated difference from published Impact on DLA estimate Impact on global estimate
Assumption of double the sampled error on excluded recently reviewed cases £5.3m 0.1 p.p 0.0 p.p
Assumption of half the sampled error on excluded recently reviewed cases -£2.6m 0.0 p.p 0.0 p.p

Cases with an upcoming review would be expected to have a higher propensity for Fraud or Claimant Error due to the length of time since their last review. By the same reasoning, cases that have had a review completed recently would be expected to have a lower propensity for Fraud or Claimant Error. Official Error is assumed to be the same regardless of recent or scheduled reviews.

Given these assumptions, and the fact that both types of reviews are present in the excluded proportion, the impact of double the sample estimate occurring in the excluded proportion is extremely unlikely and a worst-case scenario.

Underpayments

For DLA underpayments two assumptions were tested for cases excluded as recently reviewed or having a scheduled review. We have estimated the impact of the excluded cases having half the error as the sample, and as a worst-case scenario having double the error as the sampled cases. The table below shows these assumptions and their estimated difference to the published amounts and rates.

Estimated difference from published Impact on DLA estimate Impact on global estimate
Assumption of double the sampled error on excluded recently reviewed cases £12.3m 0.2 p.p 0.0 p.p
Assumption of half the sampled error on excluded recently reviewed cases -£6.2m - 0.1 p.p 0.0 p.p

These differences all fall within the published confidence intervals for the rates of overpayment and underpayment for PIP in FYE 2024, and the conclusion therefore is no adjustment is needed.

Fraud and error occurring directly on Cost of Living Payments

The estimate for Cost of Living Payments only includes fraud and error on the payments where the qualifying benefit was incorrectly paid. This means that we are omitting a small amount of fraud and error that can occur on the payment themselves. These include:

  • a small number of Cost of Living Payments which were paid in error to the wrong person
  • a number of claims that were incorrectly not paid Cost of Living Payments

The department estimates that the amount overpaid directly on Cost of Living Payments was small and well within the confidence intervals in FYE 2024. The amount underpaid directly on Cost of Living Payments would be negligible. Mop-up exercises are carried out on the cases incorrectly not paid, as soon as they are identified, and these cases are then paid at a later date.

Time Lags

The time lags involved in the fraud and error measurement process mean that further omissions are possible. Any policy or operational changes in the last five months of the financial year will not usually be covered by the reviews feeding into the publication, as the reviews tend to finish in the October of that financial year. In addition, some cases do not have a categorisation by the time the estimates are put together, often due to an ongoing fraud investigation. “Estimated outcomes” are generated for these cases for the purposes of the statistics, made by the review officer estimating the most likely outcome of the case, or based on the results from the reviews of similar cases that have been completed.

For all benefits we carried out additional work to better understand any implications of major policy/operational changes within the financial year. The conclusion was that we felt the sample period was representative of the financial year. See section 5 for further details.

Work Capability Assessment

When ESA was measured between FYE 2014 and FYE 2023, the measurement of ESA did not include a review of the Work Capability Assessment; this is also the case for the Work Capability Assessment for claimants on UC.

Universal Credit Transitional Protection

Universal Credit (UC) was introduced to replace older (legacy) benefits, including tax credits. Benefit claimants have gradually moved onto UC through:

  • natural migration – when the claimant experiences a change in circumstances, and they need to make a new claim for a benefit that UC has replaced
  • voluntary migration – when the claimant voluntarily moves to UC from their existing benefit
  • managed migration – when the claimant does not choose to migrate voluntarily and has not migrated naturally

Transitional protection can be applied to claimants who are moved onto UC through the managed migration process. A transitional protection element is applied to ensure that eligible households, with a lower calculated award in UC than their legacy benefit awards, will see no difference in their entitlement at the time they are moved to UC, providing that there is no change in their circumstances during the migration process.

The transitional protection element is calculated during the managed migration process. It is based on the circumstances for the eligible household and their legacy benefits in payment that are being replaced by UC. There is potential for the transitional protection element to be paid incorrectly if the calculation is made incorrectly and/or the legacy benefit awards in payment are incorrect based on the claimant’s circumstances. We are unable to review the transitional protection element calculation or the legacy benefit awards because these can be derived from benefits not administered by DWP. Consequently, fraud and error on this element of UC is omitted from the estimates.

Latest figures on UC Transitional Protection seem to show there are around 85,000 UC cases with this element. For these cases the transitional protection only accounts for a small fraction of the total UC award in payment. Expenditure on the transitional protection element was low in FYE 2024 compared to the total UC expenditure. Therefore, the omission was taken to be negligible, and no adjustment was made to the UC fraud and error estimates.

Knock-on impact on other benefits

We only review the benefit that has been selected for a review, and do not assess any consequential impacts on other benefits. However, in certain circumstances, for some benefits, there may be a knock-on impact on other benefits. An example of this is how changes in entitlement to DLA or PIP affects disability and carer premiums on income-related benefits (specifically IS, PC and HB), as well as CA. We account for this in our estimates by using DWP’s Policy Simulation Model to assess the impact. The Policy Simulation Model is the main micro-simulation model used by DWP to analyse policy changes and is based on the annual Family Resources Survey. However, there are other knock-on impacts which we do not cover. For example, any benefit which considers another benefit as an income. If that other benefit is incorrect then only its impact on the benefit we measured is captured. This means that potentially the global figure is an overestimation as in these scenarios the overpayment on the benefit measured would be offset by an uncaptured underpayment on the other benefit.

Third party deductions

The accuracy of third party deductions is not measured (i.e. whether the deduction is at the correct amount and is still appropriate). Third party deductions can take place to cover arrears for things like housing charges, fuel and water bills, Council Tax and child maintenance. The rate of benefit is not impacted by any third party deductions, and the amount of any Fraud or Error is based on the “gross” amount of benefit in pay.

UC sanctions

For UC, we do not assess whether the Department follows correct “labour market” procedures and takes any necessary follow up action for non-compliance by claimants (i.e. considers whether a sanction should apply if a claimant fails to apply for a job or leaves a job voluntarily). However, if a sanction decision has already been made when we review a case, then we do assess whether the impact this has on the benefit award is correct.

UC surplus self-employed profit/loss

For UC, we only measure income in the assessment period we are checking. Self-employed people must report their income on a monthly basis. If they receive income that removes their entitlement to UC in one month, and this is above the surplus earnings threshold, then any extra income is carried forward into the next month (the surplus earnings threshold is defined as £2,500 above the amount that removes their entitlement to UC in that month). If a self-employed claimant incurs a loss of any amount within an assessment period, this loss is also rolled forward to the next assessment period. When reviewing the benefit, any rolled forward income or loss is assumed to be correct.

Earnings from the hidden economy

These are claimants who are working but are not declaring those earnings to the government. For every means tested benefit, where capital and earnings affect the award, we require bank statements for all the claimant’s accounts that cover the period of the payment we are checking. This means that any earnings which goes through the claimant’s bank are likely to be picked up when those bank statements are checked.

Although we think we capture most of the Fraud related to hidden economy earnings, it is likely that not all of this would end up under the “Earnings/Employment” error reason. If the claimant fails to send in bank statements after multiple prompts, then their benefit is suspended and ultimately terminated. In these circumstances a whole award Fraud would be recorded (see Causal Link part of section 5 for more information). However, given they were hiding their earnings from the Government it is likely we would not know the underlying reason so the Fraud would be categorised as “Failure to provide evidence/engage”.

This means that the only earnings we would not pick up are those which are only “cash in hand” and are not being deposited into a bank account. We expect the impact of this to be minimal, particularly since COVID-19, many cash only businesses have diversified into accepting bank transfers/card which further reduces this omission.

Cyber-crime

We do find errors relating to this and they would be included within the “Conditions of Entitlement” error reason. The benefit reviews that underpin the statistics are very robust and encompass not only a lengthy interview with the claimant but also evidence to verify all their circumstances. Therefore, it would be very difficult for a fraudulent claimant to meet all these requirements without alerting the suspicions of the reviewing officer.

Similar to hidden economy earnings, we think we capture most of the Fraud related to cyber-crime, but it is likely that not all of this would end up under the “Conditions of Entitlement” error reason. If a claim is fraudulent then they are likely to either not attempt the interview or not provide the requested evidence, in which case a whole award Fraud would be recorded (see Causal Link part of section 5 for more information). However, it is likely that we would have no evidence as to why they did this so the Fraud would be categorised as “Failure to provide evidence/engage”.

Benefit Advances

One of the largest current omissions from our estimates is benefit advances, which are out of scope of our measurement.

UC supports those who are on a low income or out of work. It includes a monthly payment to help with living costs. If a claim is made to UC but the claimant is unable to manage financially until their first payment, they may be able to get a UC Advance, which is then deducted a bit at a time from future payments of the benefit.

The benefit review process for the fraud and error statistics examines cases where benefit is in payment. A benefit advance is not a benefit payment and is not included in the DWP expenditure figures or in our measurement process. Claimants who progress to receive payment of a benefit will be included within the scope of our measurement, but we will only review the existing benefit payment. This will not examine Fraud or Error that may have existed in any prior benefit advance payment. Claimants who only receive a benefit advance, but do not go on to receive a subsequent benefit payment, will not be included within the measurement. Advances are available for a number of benefits but, for FYE 2024, advances for UC constituted the vast majority of expenditure on benefit advances.

We estimate that for FYE 2024 the monetary value of fraud and error on UC advances lies between £0m and £60m.

Rounding policy

In the publication and reference tables, the following rounding conventions have been applied:

  • percentages are rounded to the nearest 0.1%
  • expenditure values are rounded to the nearest £100 million
  • headline monetary estimates are rounded to the nearest £10 million
  • monetary estimates for error reasons are rounded to the nearest £1 million

The proportion of claims paid incorrectly is rounded to the nearest 1% in the publication and expressed in the format “n in 100 cases”. The reference tables present the same values as a percentage rounded to the nearest 0.1%.

Individual figures have been rounded independently, so the sum of component items do not necessarily equal the totals shown.

4. Sampling and Data Collection

The fraud and error statistics are determined using a sample of benefit records, since is it not possible to review every benefit record. The sample of benefit records provide data from which inferences are made about the fraud and error levels in the whole benefit claimant population.

The number of benefit records to be reviewed is determined by a sample size calculation. The sample size calculation is used to ensure that a sufficient number of benefit records are sampled, which allows meaningful changes in the levels of fraud and error to be detected for the whole benefit claimant population.

Benefit records are selected on a monthly basis from data extracts of the administrative systems. The population from which the samples are drawn are the benefit records that are in payment in a particular assessment period, that is where there is evidence of a payment relating to the previous month. This is known as the liveload.

The monthly samples are taken from the liveload in advance of the scheduling of the benefit reviews, to give time for the sample to be checked and for background information to be gathered on each benefit record sampled. Any benefit record relating to a claimant who has been previously sampled in the last 6 months, or meets specific exclusion criteria (e.g. terminally ill) will not be reviewed.

We use Simple Random Sampling to select the sample of benefit records for each benefit that is reviewed in the current year. Benefit records are sampled randomly to ensure an equal chance of being selected for the sample. The sampling methodology is used to attempt to minimise selection bias in the sample and aims to select a sample that is representative of the entire benefit claimant case population.

The benefits sampled for this year and the methodologies applied are as follows:

Simple random sample:

  • Disability Living Allowance
  • Pension Credit
  • Universal Credit
  • State Pension
  • Personal Independence Payment

Housing benefit methodology uses simple random sampling stratified by Primary Sampling Unit (PSU) and four different client groups:

  • Working Age in receipt of IS, JSA, ESA, PC or UC
  • Working Age not in receipt of IS, JSA, ESA, PC or UC
  • Pensioners in receipt of IS, JSA, ESA, PC or UC
    • Pensioners not in receipt of IS, JSA, ESA, PC or UC

Note: for HB only the client groups “Pensioners in receipt of IS, JSA, ESA, PC or UC” and “Pensioners not in receipt of IS, JSA, ESA, PC or UC” were reviewed in FYE 2024.

Abandoned Cases

Of the benefit records sampled, there are some that are not eligible for a review according to strictly defined criteria for abandonment. Benefit records that fall into this category could include:

  • the claimant has a change of circumstances that ends their award before the interview can take place
  • the claimant has had a benefit reviewed in the last six months
  • if the claimant or their partner is terminally ill

When such cases occur in the sample, they are replaced by another case from a reserve list. However, for a small number of abandoned cases replacement is not possible for practical reasons. This occurs when cases are abandoned towards the end of the review year, which means that there is not enough time for a replacement case to complete the full process.

Abandoned Cases FYE 2024

It is the decision of the Performance Measurement (PM) team, during the preview stage of a case, if a case should be abandoned.

Table showing the main reason for cases abandoned for each benefit

Abandonment Reason / Benefit DLA HB PC PIP SP UC Total
Benefit not in payment/ceased or suspended 17 145 69 20 8 344 603
New activity within 6 months of the start of the sample 320 122 442
Sensitive issues 39 25 65 16 16 25 186
Allowable Absence 12 30 12 5 4 24 87
Corporate appointee with no named contact 6 4 18 34 1 8 71
Miscellaneous 38 71 31 60 59 138 397
Total cases abandoned by benefit 432 275 195 257 88 539 1,786
Total cases reviewed 1,385 3,001 1,989 1,395 1,557 3,986 13,313
Abandonment rate per benefit 31% 9% 10% 18% 6% 14% 13%
Percentage point change from previous year -24 2 5 0 -1 -3

For FYE 2024, the abandonment reasons ‘Operational Issues’ & ‘Incorrectly Sampled’ have been added to the category ‘Miscellaneous’. This is due to other abandonment reasons accounting for a greater proportion of cases abandoned.

The five reasons listed (excluding ‘miscellaneous’) accounted for around 78% of total abandonment, a reduction in total proportion from FYE 2023. This is mainly due to less cases being abandoned within the ‘Benefit not in payment/ceased’ category accounted for 33% of the total abandonments in FYE 2024, compared with 62% in FYE 2023. This difference mainly comes from the change in HB client group measured between years.

DLA has been reviewed for the first time this year, as such there is no previous comparison to be made for abandonments.

For HB, FYE 2024 reviewed the pension age client groups for HB as opposed to the non passported working age client group in FYE 2023. Different client groups have variation in the rate of abandonment due to the nature of the benefits with non passported working age claimants most likely to move off benefit with earnings increasing.

Below are updated descriptions for the top five abandonment reasons for FYE 2024:

  • Benefit not in Payment/ceased or suspended – this remains the largest cause of abandonment, with almost 60% of these abandonments being on UC claims. These are claims no longer in payment either because the claimant is no longer entitled to the benefit or because they are now claiming another benefit. This is primarily related to the time lag of benefit reviews commencing from sample selection period
  • New activity within 6 months of the start of the sample – This reason is solely linked to the disability benefits. This is due to actions previously done or planned to be done on claimant’s case such as renewal of the claim to DLA or the claimant notifying a change to the department before the review process has started. For PIP, this could be appeals or interventions made on their claim
  • Sensitive Issues – this is a reason that affects all benefits reviewed. The claimant/partner being terminally ill or finding out they recently became deceased or bereaved are the main causes for the usage of this reason
  • Allowable Absence – This reason is when the claimant is under hospital care during the review week, with no known indication of when they would be able to complete the review or being out of the country
  • Cooperate Appointee with no named contact – This reason has previously been included within ‘Miscellaneous’ in previous years. These claims relate to when the review has been planned and are unable to find the correct official to deal with the review on the claimant’s behalf
  • Miscellaneous – this category covers all remaining categories of abandonment used. These reasons are within the pre-defined criteria for abandonment. All the reasons here are unavoidable, out of our control or can’t be identified at the sample production stage of the process

Official Error Checking

For all benefits that are not UC the week of the official error check is changeable and is decided by either the day the claimant is notified of the review or the date the review takes place. For UC, the check is usually for the last Assessment Period (of one month) prior to the date the review takes place.

Specially trained DWP benefit review officers carry out the checks. The claimant’s case papers and DWP computer systems are checked to determine whether the claimant is receiving the correct amount of benefit according to their presented circumstances. This identifies any errors made by DWP officials in processing the claim and helps prepare for the next stage: a telephone review of circumstances with the claimant.

Claimant Error and Fraud Reviews

For all benefits, benefit review officers normally check for Claimant Error (CE) or Fraud by comparing the evidence obtained from the review to that held by the Department. The claimant may not be interviewed if:

  • the case is already under an ongoing fraud investigation
  • a suspicion of Fraud arises while trying to secure an interview

When such cases occur in the sample, the outcome of the fraud investigation is used to determine the review outcome.

Where, following receipt of a letter informing them of a review, the claimant reports a change of circumstances that results in entitlement to that benefit ending before the review takes place, an outcome of Causal Link would be considered without the claimant being interviewed. See section 5 for information on Causal Link errors.

Types of errors excluded from our estimates

Types of errors:

  • Some failures by DWP staff to follow procedures are not counted as official errors; where the failure does not have a financial impact on the benefit award or where the office have failed to take action which could have prevented a claimant error from occurring. These are called procedural errors
  • Accounting errors are errors where, despite an error in the claimant’s benefit, an overpayment (or underpayment) of the benefit undergoing a check could be offset against any corresponding underpayment (or overpayment) on the same benefit or in the case of State Pension and Pension Credit each other. For State Pension and Pension Credit these errors are excluded from the monetary estimates but are included in estimates of the proportion of claims paid incorrectly. For errors where the offset is on the same benefit they are excluded completely
  • A notional error is one where, following PM intervention, the claimant is advised to apply for State Pension or an occupational pension, but the claimant does not apply for it so the DWP decision maker assumes a ‘notional’ amount for this income based on the amount the claimant could receive if they apply for it. These notional amounts are taken into account from a future date so are excluded from our estimates

Recording Information

Case details relating to the fraud and error reviews are recorded on internal bespoke ‘fraud and error’ software (the system is known as FREDA), to create a centrally held data source. This can then be matched against our original sample population to produce a complete picture of fraud and error against review cases across our sample.

5. Measurement Calculation Methodology

Fraud and error measurement relies on three data sources:

  • Raw Sample held on ‘FREDA’ (the database on which the review outcomes are recorded), is used to identify the Monetary Value of Fraud and Error (MVFE) for individual cases, categorise its cause and quantify it as a proportion of the sample
  • Benefit Population data to estimate the extent of fraud and error across the whole claimant caseload from the sample data
  • Expenditure data to estimate the total MVFE to the Department

Estimates are categorised into overpayments (OP) and underpayments (UP) and one of three incorrectness types: Claimant Fraud (CF), Claimant Error (CE) or Official Error (OE). Further sub-categories of error reasons are used to provide more details about the nature of the Fraud or Error. Details on error classifications can be found in the glossary at Appendix 3.

Detailed below are the main calculation steps that the Fraud & Error Measurement and Analysis (FEMA) team carry out to produce the final Fraud and Error estimates.

Methodology for Benefits reviewed this year

Benefits that have been reviewed this year account for 83% of the total benefit expenditure. For each of the benefits reviewed this year a random sample of cases was taken. See section 4 for further details.

An Official Error check is carried out; see Official Error Checking part of section 4. The claimant is then contacted, and a review carried out with evidence requested to verify their circumstances as outlined in Claimant Error and Fraud Reviews part of section 4.

Finally, a case is categorised as Benefit Correct, Official Error, Claimant Error or Fraud (or a combination of the last three).

There are specific scenarios and adjustments that we then take into account. These are detailed below:

Causal Link

Cases where there is a change to the claimant’s award as a result of the review activity or, after initial contact, the claimant subsequently fails to engage in the review process, are categorised as Claimant Fraud with causal link. Action is taken to suspend their payment and subsequently terminate their claim.

Examples of behaviours that can trigger cases to be categorised as Causal Link include:

  • the claimant receives notification of the review and subsequently contacts the department to report an immediate change e.g. living with a partner, starting work, self-employment or capital changes. Then supporting evidence needed to verify the change is not provided, resulting in claim suspension and termination
  • the claimant completes the review but subsequently declares that a change has happened shortly following the period of the review
  • the claimant receives notification of the review and does not engage in the review process or contacts the department with a request to withdraw their claim
  • the claimant completes a review and declares a change, however supporting evidence needed to verify the change is not returned, resulting in claim suspension and termination

For UC there are cases where the claimant fails to engage in the review process, but there is supporting evidence that a change is not due to the review. These are categorised as ‘mitigating circumstances’. For these cases, information is available on our systems to indicate why the person may not have engaged. In most cases, they have moved into paid work following the Assessment Period under review.

For all benefits post-review, every Causal Link error is categorised as either high suspicion or low suspicion. This categorisation is used in the netting and capping procedure (see below section Netting and Capping) to help attribute losses to the error reasons we are most confident about. Any losses attributed to low suspicion Causal Link after netting and capping will have their error reason changed to “Failure to provide evidence/fully engage in the process”.

Examples of high suspicion Causal Link errors include:

  • shortly after review the claimant terminates their claim (rather than send in evidence)
  • the claimant told us at the review of a change of circumstances, but we cannot confirm that the change occurred prior to (or in) the assessment period checked
  • post-review, the claimant made a change of circumstances that cannot be confirmed as starting after the assessment period checked
  • when asked to send in more information or to clarify further queries on evidence sent in, the claimant stops engaging

The main reason that we would class a case as low suspicion is where the claimant fails to send in evidence, but we have no prior suspicions of fraudulent intent.

Adjustments

A series of adjustments are made to the sample data, to allow for various characteristics of the benefits and how their data is collected and recorded.

The following table highlights which adjustments apply to each of the benefits reviewed in FYE 2024.

SP DLA PC HB UC PIP
Netting and Capping Y Y Y Y Y Y
Estimated Outcomes Y Y Y Y Y Y
New Cases Factor N N Y Y N N
Underlying Entitlement N N N Y N N
Cannot Review Cases Y Y Y Y Y Y
Reasonably Expected to know N Y N N N Y

Netting and Capping

Where a case has more than one error, these errors can be “netted off” against one another to produce a total value. For example, if a case is found to have two different OEs, one leading to an UP and one leading to an OP, then these are “netted off” to produce a single OP or UP. This is done to better represent the total monetary loss to the public purse (via OPs) or to the claimant (via UPs).

The monetary loss on each case is the difference between the case award paid at the review/assessment period, and the correct award calculated following the review – the “award difference”.

A case may have OPs of more than one ‘type’ which sum to a total greater than the award difference. To ensure that the total OP does not exceed the total award difference, we ‘cap’ the OP amount using a hierarchy order of actual CF, Causal Link (high suspicion) CF, Causal Link (low suspicion) CF, CE then OE. This capping process means that a small proportion of CE and OE found during the survey is not included in the estimates, and therefore the final estimates may actually be under-reporting CE and OE in the benefit system, but the total amount of fraud and error is correctly reported.

Estimated outcomes

For a number of cases reviewed this year, the review process had not been completed at the time of the analysis and production of results, often due to incomplete fraud investigations. As a result, predictions for the final outcomes for these cases have been made in the analysis using either the review officer (RO) estimation of the most likely outcome, or the results from the reviews of similar cases that had been completed.

New Cases Factors

New Cases Factors are an adjustment applied to help ensure that the durations on the sample accurately reflect the duration on benefit within the population.

As a result of the time required to collect the information needed to review a case, as well as other operational considerations, there is an unavoidable delay between sample selection and case review. This delay means that fewer low duration claims will be represented in the sample of cases, which artificially introduces a bias around claim durations at the point of interview.

Cannot Review

Cases that cannot be reviewed, primarily due to the claimant not engaging in the review process resulting in their benefit claim being suspended and later terminated, are initially recorded as Fraud. These cases are referred to as ‘Cannot Review’ and for most cases the Department holds very little evidence of their current circumstances and their reasons for failing to engage.

Not all of these cases will be Fraud so for cases where there is a lack of evidence available, additional checks are conducted at a later date. These checks are to determine if the individuals have reclaimed benefit and if there was a suspicion of Fraud recorded on the case at the initial preview. The outcome of these checks will result in these cases being re-categorised for reporting purposes. 0.8% of sampled cases in FYE 2024 did not have an effective review and we had no evidence as to why.

There are three different categories that are applied to cannot review cases for reporting purposes:

  • Not Fraud – If the individual reclaims benefit within 4 months, with the same circumstances and at a similar rate they were receiving prior to review, then the Fraud is removed
  • Fraud remains – If an individual does not reclaim benefit and there was a suspicion of Fraud raised at the preview stage of the review then the case remains as Fraud
  • Inconclusive – If the individual does not reclaim benefit and there was no suspicion of Fraud at the preview stage of the review then the case is categorised as inconclusive as there is no evidence to suggest the case is Fraud or not

Inconclusive cases are excluded from the estimates and reported separately in footnotes in the publication and reference tables.

Failure to Provide Evidence

If a claimant fails to provide requested evidence after a Performance Measurement review, their benefit is suspended and then subsequently terminated. A whole award error is then recorded relating to the evidence that we failed to receive (for example, if a bank statement which was requested to verify Capital was not received, then a whole award Capital Fraud is recorded). However, if we had no real suspicions about the evidence apart from the claimant not providing it, we reclassify the whole award error into Failure to provide evidence. For more information on this please see the section on Causal Link.

We adopt a data driven approach to attempt to reclassify some of the failure to provide evidence errors into known reasons for Fraud or potentially remove the Fraud entirely. This involves looking at data we have on a case four months after classification, to see if this gives us a clearer picture of why they chose not to provide evidence. If we can determine the reason for the fraud, we reclassify the error again from Failure to Provide into the error reason the data suggests. However, if the information gives us reason to believe that the case is not fraudulent, the fraud is removed (i.e., information from HMRC shows the claimant entered full time work before the claim was terminated and their new earnings would have meant they were no longer eligible for the benefit)

Benefit-specific adjustments

This section contains details of any benefit-specific sampling issues, or things that we only do for certain benefits when we calculate our estimates.

Universal Credit

Zero payment cases: Universal Credit cases still live but with zero entitlement are not included in our sample for benefit reviews. However, when calculating the proportion of claims paid incorrectly on Universal Credit we scale the final figure to account for these cases.

State Pension

Working for a certain period in the UK means that individuals are entitled to a UK State Pension from State Pension age even if they subsequently move outside of the UK (before or after reaching State Pension age).

Reviews for State Pension only cover cases from Great Britain (GB). We apply the GB rate to the whole of the State Pension expenditure, including those cases living overseas. We also estimate an additional amount of Claimant Error for the impact of late or non-notification of death on State Pension cases living overseas. We see very low rates of Fraud and Claimant Error on the GB caseload and therefore it is reasonable to assume that, aside from late or non-notification of death, we would find equally low rates for claimants of State Pension living overseas.

The International Pension Centre collects information on deaths of overseas State Pension claimants, but do not consistently collect any information on other changes of circumstance on these cases. This means that it is only possible to measure Fraud and Error overpayments relating to late or non-notification of death.

Two methods are used by DWP to confirm that overseas SP claimants are still alive and entitled to the benefit. These are as follows:

Life Certificates

A life certificate (LC) is a paper-based form that should be completed by the claimant, signed by a witness and then returned by the claimant. If the completed form is not returned after 16 weeks, then the claimant’s benefit is suspended and another LC is sent out, then following a further 16 weeks with no response, the claimant’s benefit is terminated. If the LC is returned by the claimant, then their benefit entitlement continues (subject to any changes in rate due to changes of circumstance reported by the claimant).

Death Exchange

DWP exchanges death data with Australia, Germany, Isle of Man, Malta, Netherlands, New Zealand, Poland, and the United States of America. Most of this data is received monthly. The process for death exchanges begins with these countries requesting lists of claimants living in their country and receiving a UK State Pension. Following this, they send the death data for DWP to process.

Methodology

We estimate the additional amount of Claimant Error due to late or non-notification of death using an estimate calculated in 2006. This was based on data from the January 2004 life certification exercise and is applied to the whole of the State Pension overseas expenditure.

Life certificates were suspended during COVID, restarting in November 2021. The life certificate exercise runs over a two-year period. We are looking to update the 2006 estimate in the future, but need to ensure that the data is complete and robust before doing so.

Personal Independence Payment

For disability benefits, there are some changes which the claimant should report (for example, hospitalisation). However, many changes are gradual improvements or deteriorations in their medical needs, and it is difficult for some claimants to know at what point these needs have changed sufficiently to affect their benefit entitlement.

PIP legislation states that when a case is reassessed and their benefit is reduced, the Department will only seek to recover an overpayment when it is reasonable for the claimant to have known they should have reported the change. In other cases, the benefit will be treated as correct up to the point of reassessment.

It was identified that during PIP reviews there appeared to be variance in the application of “Reasonably expected to know” decisions, resulting in such cases not always having overpayments reported.

In Spring 2017, PM staff completed a joint exercise with PIP Operational staff to reconsider all of the information available to identify improvements to the review process to ensure the measurement of PIP was sufficiently robust.

Accordingly, error cases have been excluded from the headline overpayment estimates where the claimant could not reasonably have been expected to know they should have reported it.

Special Rules Terminally Ill (SRTI)

PIP claimants are considered to be particularly vulnerable, therefore it is not always deemed appropriate to put a claimant through the review process. This is the case for terminally ill PIP claimants.

Instead, an adjustment is made to the PIP estimates to account for their exclusion from the sample. 96% of terminally ill PIP claimants receive the highest award, with many regulations. For this reason, the fraud and error rate in this group is not considered to be the same as for the rest of the PIP population, and an adjustment is made based on this assumption.

The impact of the SRTI adjustment can be found in section 3.

Disability Living Allowance

The same rules around “Reasonably expected to know” decisions applied to PIP above are also applied to DLA.

Grossing

Grossing is the term used to describe the creation of population estimates from the sample data; sample results are scaled up to be representative of the whole population.

Example of a simple grossing factor ‘G’, if we were to sample 100 cases from a population of 1,000:

G = N ÷ n
= 1000 ÷ 100
= 10

Where ‘N’ is the population or sampling frame from which the cases are selected and ‘n’ is the sample size taken.

The above grossing factor shows that, in this example (sampling 100 cases from a population of 1,000), then each case would have a grossing factor of 10 (i.e. each sample case represents 10 cases from the population). Hence if a case was shown to be in error, this would represent 10 errors once grossed.

UC is replacing a selection of legacy benefits. As this process continues the UC caseload will increase whilst those other benefit caseloads will decrease. As a result, grossing for all benefits are calculated on a monthly basis. This ensures that an error identified at the start of the year is grossed up by less than an error identified at the end of the year if the caseload is increasing (or vice versa if it is decreasing). Grossing factors are different for each benefit due to the sample, population and adjustments made.

Percentage overpaid and underpaid

The grossing factors are then applied to the sample data to calculate values for the grossed awards, the grossed overpayments and grossed underpayments i.e. these are scaled up proportionally to what we would expect to find in the population. In turn, the resulting grossed values are used to calculate the total (global) annual percentage overpaid and underpaid.

Extrapolation

The grossed results provide a core estimate of levels of fraud and error. Extrapolation aligns the monetary amount with the benefit expenditure, which is particularly important given the sample period and the financial year do not fully align.

Monetary Value of Fraud and Error (MVFE)

To then calculate the MVFE across the benefits, we apply the OP or UP percentage rates to the total annual expenditure for each benefit. This means that the MVFE is affected by the increases and decreases in expenditure, even if the OP and UP percentages are stable. We see the impacts of this in our estimates for benefits not reviewed in the current year, where we use the same rate of fraud and error from previous years but apply it to the expenditure on the benefit in the current year (which will have changed from the year before).

Although expressing fraud and error in monetary terms (i.e., MVFE terms) might be helpful for a reader to contextualize the figures, we recommend making comparisons on a year-on-year basis based on the percentage rates of fraud and error. This is particularly important for benefits where the expenditure changes a lot each year, as comparisons of monetary amounts can be misleading. For example, on a benefit with growing expenditure, it can be possible for the monetary amount of fraud and error to increase, even if the percentage rate of fraud and error has actually gone down.

Central Estimates and Confidence Intervals

The central estimates produced following extrapolation are based on reviews of random samples and hence are subject to variability. Therefore, confidence intervals are provided with the central estimates to quantify the uncertainty associated with these estimates.

The central estimates and confidence intervals are incorporated into the Global (overall) Estimates of fraud and error. These combine all separate DWP benefits to calculate an overarching set of Fraud, Claimant Error and Official Error rates for overpayments and underpayments. See section 6 on Measurement of Total Overpayment and Underpayments for more detail.

Total Overpayment and Underpayments

The fraud and error estimates need to include all expenditure on benefits by DWP. Some benefits have been reviewed for fraud and error in the current year, and some have been reviewed in previous years. We also need to include estimates for benefits which have never been reviewed. A full list of which benefits are in scope for each release of the Fraud and Error in the Benefit system estimates is included within Appendix 2 of this document.

We also have an estimate of Interdependencies, the knock-on effect of Disability Living Allowance (DLA) fraud and error on other benefits, where receipt of DLA or PIP is a qualifying condition. This is only included within the overpayments calculation and not the underpayments.

Methodology for Benefits previously reviewed

Some benefits which were not measured this year were measured in previous years. For these benefits we apply the rate from the last time the benefit was measured to the current year’s expenditure, to get an estimate of the monetary value of fraud and error.

Benefits that have been previously reviewed account for 11% of the total benefit expenditure.

Methodology for Benefits never reviewed

As mentioned in section 2 “Changes”, we use multi criteria decision analysis to choose which benefits we should measure each year. Some benefits have a small amount of expenditure and therefore are unlikely to ever be selected for measurement. For each of the benefits that have never been reviewed, we use a similar benefit’s rate of fraud and error as a proxy. We then apply that to the expenditure on that unreviewed benefit to get an estimate of the monetary value of fraud and error.

For more information on the proxies used, please see the “Benefits never reviewed” section of Appendix 2.

Benefits that have never been reviewed account for 6% of the total benefit expenditure.

Central Estimate and Confidence Intervals

The percentage estimate (i.e., the overall rate of fraud and error) is the sum of the monetary value of fraud and error for all benefits reviewed this year, those reviewed in previous years, those never reviewed and interdependencies, divided by the overall expenditure. This is done independently for Fraud, Claimant Error, Official Error and the overall fraud and error.

The central estimate is the estimate obtained from the sample data. It provides our best guess of the unknown value that we are trying to measure.

Confidence intervals are calculated for the percentage estimates, to quantify the statistical uncertainty associated with the central estimate. Some adjustments are made to the individual benefits before this is calculated for overall fraud and error:

  • Confidence intervals for benefits reviewed previously are deliberately widened
  • Confidence intervals may be widened further if it is believed that non-sample error could impact the accuracy of the estimates. For more information on non-sample error please see Section 6 “Accuracy and Reliability”
  • Confidence intervals are widened for the benefits never reviewed, whereby the standard error is assumed to be 40% of the central estimate, to reflect greater uncertainty given the less robust method of estimation for these benefits

The uncertainty surrounding a central estimate is associated with both the variance of the outcome within the sample and the size of the sample from which it is calculated. A 95% confidence interval is used to indicate the level of uncertainty. It shows the range of values within which we would expect the true value of the estimate to lie. A wider range for the confidence interval implies greater uncertainty in the estimate.

Confidence intervals are calculated using a statistical technique called bootstrapping. It is used to approximate the sampling distribution for the central estimate. The sampling distribution describes the range of possible values, for the central estimate, that could occur if different random samples had been used.

Bootstrapping is a computationally intensive technique that simulates resampling. A computer program is used to take 4,000 resamples with replacement, of equal size, from the initial sample data. The percentage rate of fraud and error is calculated for each of the resamples. These estimates are ordered from smallest to largest and this gives the approximated sampling distribution.

The 95% confidence intervals are obtained from the bootstrapping results, by taking the 100th estimate (2.5th percentile) and the 3,900th estimate (97.5th percentile). We also check the median estimate (50th percentile) against our actual central estimate to ensure that no bias exists.

Calculation of Net Loss estimate

Recoveries refer to money recovered in the same financial year as the overpayment estimates, regardless of the period the debt is from. They include debt recovered by both the Department and Local Authorities (who administer Housing Benefit payments). The recovery data for Housing Benefit covers the period October 2022 to September 2023, due to a time lag on the data being available.

Net loss is calculated as a monetary amount by subtracting the value of the recoveries from the value of the overpayments. The percentage estimate is then calculated by dividing the monetary net loss by the expenditure. Net loss can only be calculated at the overall fraud and error level because error classification differs between overpayments and recoveries.

As the recoveries are actual values rather than estimates, the calculation of net loss does not affect the uncertainty around the overpayment estimates. The confidence intervals for net loss are calculated by subtracting the value of the recoveries from the upper and lower confidence limits. Percentage confidence intervals are obtained by dividing by the expenditure.

Net loss is calculated overall and individually for benefits reviewed this year and previously reviewed benefits. Net loss is also calculated for the group of benefits never reviewed combined, by subtracting all recoveries relating to these benefits from the total overpayments.

Some recoveries have no associated overpayments for the same period, as these benefits are no longer administered by the Department. This is because the debt relates to expenditure from previous years. In addition, some recoveries the Department makes are not included in our net loss estimate as they do not relate to our fraud and error reporting. For example, recoveries of tax credits and of benefit advances (which are outside the scope of our measurement; see section 3 for more details).

The overall net loss estimate includes the benefits reviewed this year, benefits reviewed in previous years, benefits never reviewed and recoveries for which there is no overpayment.

6. Quality Report

Introduction

This section of the document assesses the quality of the fraud and error in the benefit system national statistics using the European Statistics System Quality Assurance Framework. This is the method recommended by the Government Statistical Service Quality Strategy. Statistics are of good quality when they are fit for their intended use.

The European Statistics System Quality Assurance Framework measures the quality of statistical outputs against the dimensions of:

  • relevance
  • accuracy and reliability
  • timeliness and punctuality
  • comparability and coherence
  • accessibility and clarity

The Government Statistical Service also recommends assessment against 3 other principles in the European Statistics System Quality Assurance Framework. These are:

  • trade-offs between output quality and components
  • balance between performance, cost and respondent burden
  • confidentiality, transparency and security

These dimensions and principles cross the three pillars of trustworthiness, quality and value in the Code of Practice for Statistics.

Relevance

Relevance is the degree to which statistics meet the current and potential needs of users.

The Department for Work and Pensions (DWP) fraud and error in the benefit system statistics provide estimates of fraud and error for benefits administered by the DWP and local authorities.

The series has been developed to provide information to various users for policy development, monitoring and accountability, as well as providing academics, journalists and the general public, data to aid informed public debate.

The statistics:

  • include DWP benefits and those administered by local authorities
  • are the primary DWP indicator for levels of fraud and error in the benefit system
  • are in the DWP business plan
  • are important for DWP assurance on the impact of anti-fraud and error activity across the business

The publication is essential for providing our stakeholders with:

  • a consistent time series for assessing fraud and error trends over time
  • data to assess current DWP fraud and error policy and evaluate recent changes to these or business processes
  • the evidence base for assessing the potential effect of future fraud and error policy options and programmes
  • robust data to inform future measurement options
  • estimates of fraud and error for the DWP annual report and accounts
  • data to measure government performance relating to objective 5 of the DWP single departmental plan: transform our services and work with the devolved administrations to deliver an effective welfare system for citizens when they need it while reducing costs, and achieving value for money for taxpayers. Read the latest plan (correct at the time of publication of this document, May 2024)
  • estimates that feed into the annual HM Revenue and Customs National Insurance Fund Accounts

We recognise that our users will have different needs and we use a range of different methods to contact them. We frequently meet internal DWP users to discuss their requirements. As for external stakeholders, we often contact the National Audit Office and we occasionally contact HM Revenue and Customs and Cabinet Office.

Engagement with other external users is usually through the DWP statistical pages of this website where we:

Accuracy and Reliability

Accuracy is the closeness between an estimated result and the unknown true value. Reliability is the closeness of early estimates to subsequent estimated values.

The statistics are calculated from the results of a survey sample, which are recorded on an internal DWP database. The survey combines data collated from DWP administrative systems and local authority owned Housing Benefit systems, with data collected from the claimant during an interview.

The estimates obtained are subject to various sources of error, which can affect their accuracy. Both sampling and non-sampling error are considered in producing the statistics.

Sampling error arises because the statistics are based on a survey sample. The survey data is used to make conclusions about the whole benefit caseload. Sampling error relates to the fact that if a different sample was chosen, it would give different sample estimates. The range of these different sample estimates expresses the sample variability. Confidence intervals are calculated to indicate the variability for each of the estimates. More detail on central estimates and confidence intervals is provided in section 3.

Sources of non-sampling error are difficult to measure. However, where possible, these uncertainties have been quantified and combined with the sampling uncertainties, to produce the final estimates. Quality assurance processes are undertaken to mitigate against particular types of error (for example, data entry error).

Possible sources of non-sampling error that may occur in the production of the fraud and error statistics include:

  • Data entry error – the survey data is recorded on a database by DWP staff. Data may be transcribed incorrectly, be incomplete, or entered in a format that cannot be processed. This is minimised by internal validation checks incorporated into the database, which can prevent entry of incorrect data and warn staff when an unusual value has been input. Analysts undertake further data consistency checks that are not covered by the internal database validations
  • Measurement error – the survey data collected from the benefit reviews are used to categorise an outcome for each case. The correct categorisation is not always obvious and this can be recorded incorrectly, particularly for complex cases. To reduce any inaccuracies, a team of expert checkers reassess a selection of completed cases before any statistical analysis is carried out. This evidence is used as a feedback mechanism for the survey sample staff and also for the statistical analysis
  • Processing error – errors can occur during processing that are caused by a data or programming error. This can be detected by a set of detailed quality assurance steps that are completed at the end of each processing stage. Outputs are compared at each stage to identify any unexpected results, which can then be rectified
  • Non-response error – missing or incomplete data can arise during the survey. Supporting evidence to complete the benefit review may not be provided, or the claimant may not engage in the review process altogether. In other cases, the benefit review may not have been completed in time for the analysis and production of results. An outcome is imputed or estimated in these cases, by different methods that are detailed in this document
  • Coverage error – not all of the benefit caseload can be captured by the sampling process. There is a delay between the sample selection and the claimant interview, and also a delay due to the processing of new benefit claims, which excludes the newest cases from being reviewed. An adjustment is applied to ensure that the duration of benefit claims within the sample accurately reflects the durations within the whole caseload

The list above is not exhaustive and there are further uncertainties that occur due to assumptions made when using older measurements for benefits that have not been reviewed this year. There are also some benefit-specific adjustments that are part of the data processing.

More detailed information about the quality of the statistics can be found earlier in this document. This includes discussion of the limitations of the statistics, possible sources of bias and error, and elements of fraud and error that are omitted from the estimates.

Timeliness and Punctuality

Timeliness refers to the time gap between the publication date and the reference period for the statistics. Punctuality is the time lag between the actual and planned dates of publication for the statistics.

The fraud and error in the benefit system report is usually published around 8 months after the main reference period.

Due to the time taken to undertake the interviews and gather follow up information, final data from the reviews is not made available to analysts until 4-5 months after the start date of the last interviews. The production of the statistics and reference tables, and the associated clearance processes, then takes the analytical team about 2 months to prepare. Improvements over the last few years have reduced this from 3 months.

DWP pre-announce the date of release of the fraud and error in the benefit system report 4 weeks in advance on this website and the UK Statistics Authority publication hub, in accordance with the Code of Practice for Statistics.

The statistics are published at 9.30am on the day that is pre-announced. The online release calendar is updated at the earliest opportunity to inform users of any change to the date of the statistical release and will include a reason for the change. All statistics will be published in compliance with the release policies in the Code of Practice for Statistics.

Comparability and Coherence

Comparability is the degree to which data can be compared over time, region or another domain. Coherence is the degree to which the statistical processes that generate two or more outputs use the same concepts and harmonised methods.

Our publication provides information on the estimates over time. Where breaks in the statistical time series are unavoidable, users are informed within the report by a text explanation, with clear sectioning within the time series reference tables and detailed footnotes.

Any changes made to the DWP or local authority administrative system data are assessed in terms of their impact on fraud, error and debt strategy and policy. These are then impacted against the fraud and error measurement review process and communicated to our internal users and the National Audit Office through our change of methodology log. The same is true for any changes made to business guidance, processes and review methodology, as well as our own calculation methodology.

We agree some methodology changes in advance with internal stakeholders using change request and change notification procedures.

External users are notified of any changes to methodology in the ‘Methodology Changes’ section of the fraud and error in the benefit system report. Substantial changes to the report structure or content will be announced in advance on the fraud and error in the benefit system collection.

The fraud and error in the benefit system statistics form the definitive set of estimates for Great Britain. They are underpinned by reviews of benefit claimants in England, Wales and Scotland.

The benefit expenditure figures used in the publication also include people resident overseas who are receiving United Kingdom benefits, except for Financial Assistance Scheme payments, which also cover Northern Ireland. All other benefit expenditure on residents of Northern Ireland is the responsibility of the Northern Ireland Executive. The benefit expenditure figures do not include amounts devolved to Scottish Government (which totalled £4.4 billion in FYE 2024. Reporting the levels of fraud and error of this benefit expenditure is the responsibility of Social Security Scotland. Their estimates for FYE 2023 were published as part of their annual report.

Northern Ireland fraud and error statistics are comparable to the Great Britain statistics within this report, as their approach to collecting the measurement survey data, and calculating the estimates and confidence intervals is very similar. Northern Ireland fraud and error in the benefit system high level statistics are published within the Department for Communities annual reports.

HM Revenue and Customs produce statistics on error and fraud in Tax Credits. Again, these estimates can be compared to form a whole benefit view.

Accessibility and Clarity

Accessibility is the ease with which users can access the statistics and data. It is also about the format in which data are available and the availability of supporting information. Clarity refers to the quality and sufficiency of the commentary, illustrations, accompanying advice and technical details.

The reports and reference tables can be accessed on the statistics pages on this website and the UK Statistics Authority publication hub.

Fraud and error in the benefit system statistics follow best practice and guidance from the Government Digital Service and Government Statistical Service, in publishing statistics that give equality of access to all users.

For data protection reasons, the underlying datasets are not available outside DWP. However, the reference tables published alongside the report provide detailed estimates, giving a breakdown of overpayments and underpayments into the different types of fraud and error, for the benefits measured in that year. The reference tables are available in both standard and accessible formats.

Technical language is avoided where possible within the report. To help users, the report contains definitions of key terms such as: Fraud, Official Error and Claimant Error. A more extensive glossary of terms and error types is included in Appendix 1.

Contact details are provided for further information on the statistics, guidance on using the statistics, data sources, coverage, data limitations and other necessary relevant information to enable users of the data to interpret and apply the statistics correctly.

Trade-offs Between Output Quality and Components

Trade-offs are the extent to which different dimensions of quality are balanced against each other. The main trade-off for these statistics is timeliness against accuracy. We always assess the right balance taking into account fitness for purpose, and fully explaining any compromises in accuracy for improved timeliness.

As detailed in the timeliness and punctuality section we wait a considerable amount of time for the data to be as complete as possible before our publication process begins, to ensure that the estimates are based on data which is as final and robust as possible. This means that we usually publish data around 8 months after the main reference period.

Balance Between Performance, Cost and Respondent Burden

The DWP fraud and error in the benefit system statistics are produced from survey data which have a high respondent burden. A compulsory interview, lasting between approximately 30 minutes and 2 hours, is required for all cases sampled for claimant error and fraud checking.

The total DWP cost for production of these statistics is approximately 150 staff (full-time equivalent).

DWP are continuously looking at more cost effective and efficient options for sourcing and collecting data, reducing the burden on the respondent and the production of the estimates.

Confidentiality, Transparency and Security

All of our data is handled, stored and accessed in a manner which complies with Government and Departmental standards regarding security and confidentiality, and fully meets the requirements of the Data Protection Act (2018).

Access to this data is controlled by a system of passwords and strict business need access control. Any revisions to our publications are handled in accordance with the department’s revisions policy.

7. Future Reporting

The future coverage and scope of the national statistics “Fraud and Error in the Benefit System” is kept under review and users are kept informed of our plans via our Publication Strategy document.

Appendix 1: Glossary of abbreviations

Abbreviation Term
AA Attendance Allowance
CA Carer’s Allowance
CE Claimant Error
CF Claimant Fraud
DLA Disability Living Allowance
DWP Department for Work and Pensions
ESA Employment and Support Allowance
FEMA Fraud and Error Measurement and Analysis
FYE Financial Year Ending
HB Housing Benefit
HMRC His Majesty’s Revenue and Customs
IB Incapacity Benefit
IS Income Support
JSA Jobseeker’s Allowance
LA Local Authority
MVFE Monetary Value of Fraud and Error
OE Official Error
OP Overpayment
PC Pension Credit
PIP Personal Independence Payment
PM Performance Measurement team
PSU Primary Sampling Unit
SP State Pension
UC Universal Credit
UP Underpayment

Appendix 2: List of benefits included in fraud and error estimates

Benefits reviewed this year

Universal Credit
State Pension
Housing Benefit
Disability Living Allowance
Pension Credit
Personal Independence Payment

Benefits reviewed previously

Employment Support Allowance (FYE 2023)
Attendance Allowance (last reviewed FYE 2022)
Carer’s Allowance (last reviewed FYE 2020)

Housing Benefit:

  • passported working age (last reviewed FYE 2019)
  • non-passported working age (last reviewed FYE 2023)

Jobseeker’s Allowance (last reviewed FYE 2019)
Income Support (last reviewed FYE 2015)
Incapacity Benefit (last reviewed FYE 2011)

Benefits never reviewed

Maternity Allowance – proxy measure: Employment and Support Allowance rates relating to Abroad, Conditions of Entitlement, Earnings and Contributions only.

Severe Disablement Allowance – proxy measure: Employment and Support Allowance rates relating to Abroad, Conditions of Entitlement, Earnings and Contributions only.

Financial Assistance Scheme – proxy measure: State Pension Official Error only.

Industrial Death Benefit – proxy measure: Pension Credit Living Together rate only.

Winter Fuel Payments – proxy measure: Service Centre Measurement carried out in FYE 2023.

State Pension Transfers – proxy measure: State Pension.

Cold Weather Payments – proxy measure: Rate of whole award errors found on the last measurement of Employment and Support Allowance, Income Support, Jobseeker’s Allowance, Pension Credit and Universal Credit.

Widow’s Benefit / Bereavement Benefit – proxy measure: Employment and Support Allowance Contributory only element and rates relating to Conditions of Entitlement, lack of National Insurance Contributions or incorrect recording of National Insurance Contributions only.

Industrial Disablement Benefit – proxy measure: Personal Independence Payments.

Armed Forces Independence Payment – proxy measure: Personal Independence. Payments

Christmas Bonus – proxy measure: Rate of whole award errors found on the last measurement of Attendance Allowance, Carer’s Allowance, Employment and Support Allowance, Pension Credit, Personal Independence Payments and State Pension.

Cost of Living Payments – proxy measure: Rate of whole award errors found on the last measurement of Universal Credit, Income-related Employment and Support Allowance, Pension Credit, Personal Independence Payments, Attendance Allowance, Attendance Allowance, Disability Living Allowance, Winter Fuel Payments.

Statutory Sick Pay – proxy measure: No fraud and error

Statutory Maternity Pay – proxy measure: No fraud and error

Appendix 3: Further information on types of errors reported

The definitions of the key terms of Fraud, Claimant Error and Official Error are included at the start of this document. This section includes additional information on how we classify errors, including a detailed list of the types of errors we report for benefits reviewed in the current year.

Note that our methodology states that all errors (Fraud, Claimant Error and Official Error) found on a case are recorded separately and the full values of each error are recorded in isolation of one another. This can lead to the sum of the error values being higher than the benefit award. In such cases a capping calculation is performed (using a Fraud, Claimant Error, Official Error hierarchy) to ensure that the sum of the errors does not exceed the award, so that the monetary value of fraud and error is not over-reported. This can lead to some of the originally captured Fraud, Claimant Error and Official Error raw sample values being reduced during the calculation of the estimates.

In addition, it should be noted that an error which is initially categorised as claimant error, will instead be categorised as official error where the business unit within the Department/LA is in possession, from whatever source, of the true facts, regardless of whether the information has been processed by that business unit.

A Glossary of the current error types for overpayments and underpayments is given below:

  • Abroad – claimant left Great Britain after claim began, did not notify DWP before leaving and is confirmed to be abroad for a period that exceeds any allowable absence limit. State Pension abroad errors are different in that being abroad does not remove their entitlement to the benefit, however the uprating of State Pension differs depending on what country you reside in
  • Award Determination – where a Case Manager from the DWP makes an incorrect award of PIP, DLA or AA based on the declared functional needs of the claimant. This includes failing to consider the qualifying period
  • Capital – concealed or incorrect declaration of the amount of savings in bank or building society accounts, cash, ISA/PEPs, premium bonds, other property interests or shares that exceed the minimum value for capital limits Capital official errors include incorrect calculation by DWP staff of the value of declared savings, money and other financial assets available to the claimant, or failure to correctly adjust tariff levels and amend the benefit entitlement due
  • Childcare Costs – Childcare costs incorrectly included or excluded or an incorrect declaration of the amount of childcare costs for the childcare element of UC
  • Conditions of entitlement – undeclared changes in the personal circumstances of a claimant or their partner, that would end entitlement to a benefit. Examples are being in full-time education, long-term hospitalisation, imprisonment, death and assuming a false identity. Includes staff failing to act on information received raising doubt on basic entitlement to benefit. For Universal Credit this includes the agent accepting the claimant commitment on behalf of a claimant who has the capacity required to accept their claimant commitment
  • Contributions – errors where the National Insurance record is incorrect, including where HMRC has failed to record, or incorrectly recorded Child Benefit within the record. Additionally, errors caused by failure by DWP to action a change to the award of State Pension following receipt of information from HMRC
  • Control Activities are not carried out appropriately – failure of staff to conduct actions at the due time which otherwise may have changed the level of benefit payable. For example, failure to review a Pension Credit claim at the end of an Assessed Income Period, not conducting routine interviews where claimant non-participation can result in a benefit sanction or late notification of benefit disallowances. For State Pension these errors generally relate to failure to take action at age related trigger points or failure to action a change in the claimants marital status
  • Earnings/Employment – concealment or under-declaration of full or part-time work undertaken during the claim by the claimant or their partner. This work can be for an employer or self-employment. Staff failing to correctly calculate the amount of monthly benefit due for claimants who have declared any paid work they or their partner have undertaken during the claim
  • Element/Premium/Components – Elements (UC only) – The award of UC is made up of a number of different Elements. Some of these are treated in this report in a similar way to Premiums in Legacy benefits, for example Carer Element, Disabled Child Element and Work Capability Elements. Child Care Costs and Housing Costs (also Elements in their own right) are recorded separately

Premium: DWP can pay additional amounts in means-tested benefits when other benefits are also being paid, for example Disability or Carers Allowance. This often introduces additional criteria for staff to consider before deciding the qualification for the extra amounts.

Components (SP only): The award of SP is made up of a number of different Components (for example Graduated Pension, State Second Pension, Additional Pension etc). This often introduces additional criteria for staff to consider before deciding the qualification for the additional amounts.

  • Failure to provide evidence/Fully engage in the process – where the reason for the error is unclear. We are confident that the case is fraudulent, as the claimant has forgone their right to benefit, but with the evidence we have available, we cannot be certain as to why
  • Functional Needs – where the claimant has failed to declare a change in their ability to carry out any of the activities on which PIP or AA is considered, or misrepresented their abilities when making their claim (whether intentionally or not)

Hospital/Registered Care HomeAA, PIP and DLA (excluding DLA children) is not normally payable for any period or periods of more than 28 days, during which a customer is being maintained free of charge whilst undergoing treatment as an in-patient in a hospital or similar institution or is resident in a care home where the Local Authority (LA) meets the costs of any of the qualifying services.

For UC couple cases, a person who is living away from their partner ceases to be treated as a member of a couple and part of the same benefit unit, where they are absent from the household, or expected to be absent from the household, for 6 months or more.

  • Household composition – failing to disclose changes in household composition, for example a non-dependant leaving. Claiming incorrectly for children which increases the value of DWP benefits payable, or claiming for adult partners who leave the household

Includes incorrect action by staff, taken in respect of other people the claimant declared living with them. This includes a partner or dependent children incorrectly omitted or included in the assessment (with due regard to entitlement to any disability premium or benefit reductions due to long-term hospitalisation).

  • Housing costsDWP provides financial assistance with the costs of ground rent and service charges. Claimants fail to disclose payment of changes to housing costs. For example service charges, or fail to declare sale of property, or a change of address which would end the extra amounts payable. For Universal Credit this includes determining the housing cost element that replaces housing benefit, this can include calculation of the rent amount, correctly incorporating rent free weeks and size criteria. DWP can also inaccurately calculate the ground rent or service charges

  • Income - Occupational and Personal Pensions – concealed or incorrect declaration of income received from a non-state pension, obtained through contributions paid in past employment schemes, annuities or personal investments

DWP staff failing to take into account the correct amount of non-state pensions declared by the claimant.

  • Income - Other – concealed or under-declaration of income coming into the household, from sources such as sick pay from work, spousal maintenance, partner’s student income, unemployment or similar insurance policy payments

Failure by staff to correctly identify or record other money coming into the household, such as Child Benefit, sick pay from work, spousal maintenance, partner’s student income, unemployment or similar insurance policy payments.

  • Income - Other benefits – concealed or under-declaration of income received from another benefit

DWP IT systems or staff have failed to take into account the correct value of other social security benefits currently paid to the claimant or partner, including benefits paid by a foreign state.

  • Living Together – where a claimant declares to be single but has failed to declare they actually live with another person and maintain a joint household
  • Other – this covers a range of different cases not covered in the categories above or below
  • Passporting – relevant to Housing Benefit only. It includes communication failures between different IT systems that notify the termination of a claimant’s means-tested DWP benefit to the Local Authority, impacting the HB award, or LA staff failing to act on the information received
  • Residency – errors relating to Housing Benefit claimants only, where DWP confirms that the claimant no longer lives at the address being paid for. For Universal Credit, where the claimant no longer lives at the address for which they are being paid Housing Costs Element, the error would appear in Housing Costs.
  • Tax Credits – errors where the amount or existence of tax credits results in an incorrect award of a DWP benefit
  • Uprating – errors where the IT system that pays State Pension has applied an annual increase to an element of State Pension incorrectly

Appendix 4: Glossary of Statistical terms

Key statistical terms used in this report are explained below.

  • 95% Confidence Interval: The range of values within which we can be 95% sure that the true value we are trying to estimate lies. It is used as a measure of the statistical uncertainty in an estimate
  • Estimate: An estimate is an indication of the value of an unknown quantity based on observed data. It provides information about unknown values in the population that we are trying to measure
  • Population: A population is any entire collection of items from which we may collect data. It is the entire group that we are interested in, which we wish to describe or to draw conclusions about (generally benefit claimants or expenditure in the context of this report). There are two different types of population referred to in sampling:

1. The Target Population - consists of the group of population units from whom we would like to collect data (e.g. all people claiming a benefit)

2. The Survey Population - consists of the group of population units from whom we can collect data (e.g. all claimants with sufficient case details on our datasets). The Survey Population is sometimes referred to as the ‘Sampling Frame’

  • Sample: A group selected (randomly in the context of this report) from a larger group (known as the ‘population’). Through analysing the sample, we aim to draw valid conclusions about the larger group

Appendix 5: List of historical methodology changes

Below is a list of the historical methodology changes that have been made since FYE 2006.

Methodology Change Included in which published report
Changes to sampling and calculation methods were aimed primarily at making the estimates better represent overpayments in the whole of IS, JSA and PC expenditure. Fraud and Error in the Benefit System April 2005 to March 2006 Spending Review 2004 target baseline
Definitional changes to what is being measured were introduced primarily to make the estimates better relate to the actual impact of fraud and error on expenditure. Fraud and Error in the Benefit System: April 2008 – March 2009 – Revised Edition
Incapacity Benefit started to be continuously reviewed for fraud and claimant error as well as official error. Fraud and Error in the Benefit System: April 2008 – March 2009 – Revised Edition
There was a change in this report to the calculation of the sample weightings for Income Support, Jobseeker’s Allowance and Pension Credit. Fraud and Error in the Benefit System: April 2009 – March 2010
During 2011 the Fraud and Error Measurement (FEM) team have carried out a series of changes to the calculation processes and methodology in order to simplify and align these across the individual benefits. This work has made our processing quicker, more efficient, robust and transparent and easier to quality assure. This in turn will reduce risk in our calculation processes and will enable our team, in the future, to make changes to the computer programs more easily, especially with the advent of Universal Credit, and be more flexible with resources. The new aligned processing system will also make it easier for our customers to interpret and compare findings across the individual benefits. Fraud and Error in the Benefit System: FYE 2011 Estimates
We have introduced a change to the order in which Income Support, Jobseeker’s Allowance and Pension Credit errors are capped within our calculation methodology. They are now capped for fraud first, then claimant error, then official error. In previous reports they were capped for official error first, then fraud, then claimant error. This change aligns the above benefits with the Housing Benefit capping hierarchy. Fraud and Error in the Benefit System: FYE 2011 Estimates
Since the introduction of Employment and Support Allowance in October 2008 no new claimants have been awarded Incapacity Benefit as they claim Employment and Support Allowance instead. In addition, all current claimants of Incapacity Benefit are being reassessed and will be moved to either Employment and Support Allowance or Jobseeker’s Allowance in the near future. For this reason, we have stopped measuring Incapacity Benefit for fraud and error on a continuous basis and have re-deployed resources to measure Employment and Support Allowance for Official Error instead, which was reported for the first time in May 2013. For this report and after, the preliminary 2010/11 Incapacity Benefit estimates will be used in our publications and applied to the latest expenditure figures to provide the most up-to-date monetary values of fraud and error. Fraud and Error in the Benefit System: FYE 2011 Estimates
A new error code framework was introduced in April 2010 following internal stakeholder consultation and agreement to have more meaningful information on the types of fraud and error. The 2010/11 publication was the first report to include these error code breakdowns. The breakdowns are not comparable to previously published error code breakdowns, i.e. reports before 2010/11. For this report and after, the preliminary 2010/11 Incapacity Benefit estimates will be used in our publications and applied to the latest expenditure figures to provide the most up-to-date monetary values of fraud and error.
A change was made to the way in which the extrapolation adjustment was being calculated for Income Support, Jobseeker’s Allowance and Pension Credit to ensure it was based on up-to-date assumptions for these benefits. This change was introduced from the Preliminary 2011/12 report onwards. Fraud and Error in the Benefit System: FYE 2012 Estimates
Removal of stratifications for Pension Credit by age (over 80 and under 80): This is a sampling change that was implemented from October 2012; therefore, this is the first set of statistics that is partly affected by this change. Fraud and Error in the Benefit System: FYE 2013 Estimates
Change to significance testing for benefits reviewed this year: Improved methodology where we use the bootstrapped values of the estimates of both years. We calculate the difference between each of the bootstrapped values and calculate the 95% confidence interval around the mean. If this confidence interval does not straddle zero, the change is marked as “statistically significant”. Fraud and Error in the Benefit System: FYE 2013 Estimates
Employment and Support Allowance estimates of fraud and claimant error are included in the “Continuously Reviewed” estimates for the first time in the Preliminary 2013/14 results. Full reviews began in October 2012. There is an impact on the Global Estimates as previously fraud and claimant error were estimated using a proxy measure, combining both Incapacity Benefit and Income Support Disabled and Others results. Fraud and Error in the Benefit System: FYE 2014 Estimates
An additional level of stratification was introduced into the Pension Credit single review sampling from April 2013. The new classification is based on characteristics of the claim, as held in our administrative records. This modification will not lead to any systematic change in our central estimates, but was introduced to reduce the width of our confidence intervals, thereby improving the precision of our central estimates. Fraud and Error in the Benefit System: FYE 2014 Estimates
A level of stratification was removed from the Income Support sampling from October 2013. Results can no longer be presented as split by “Lone Parents” and “Disabled/Other”. This modification will not lead to any systematic change in our central estimates, but will increase the width of our confidence intervals, thereby reducing the precision of our central estimates. The reason for the change was the reduction in caseload. Fraud and Error in the Benefit System: FYE 2014 Estimates
Half of the Income Support single review cases were randomly allocated to receive notified visits from October 2013 onwards. This has had no significant effect on the rates of fraud and error reported. Fraud and Error in the Benefit System: FYE 2014 Estimates
Housing Benefit measurement methodology changed for the 2014/15 preliminary results. We have aligned the treatments of cases with both overpayments and underpayments across Housing Benefit and the other continuously reviewed benefits in the May 2015 release. This means that both the overpayments amounts and the underpayments amounts have been reduced for Housing Benefit and hence for the total of all benefits. We have also reduced the amount of Claimant Untraceable fraud and error that we count. Fraud and Error in the Benefit System: FYE 2015 Estimates
Introduced a new ‘Global Net Loss’ measure that takes away the overpayments that the department and Local Authorities recover from the overpayments. Instead of counting all overpayments as a loss to the system, we subtract the amount that the department gets back, giving a more accurate representation of the cash loss to the public purse. Fraud and Error in the Benefit System: FYE 2015 Estimates
From 2014/15, within the measurement system, there was a change in the way some errors were classified as either claimant error or fraud, following a review of the evidence gathering process by the Performance Measurement teams. The outcome from the review emphasised the need for further questioning to establish the facts around any changes in circumstances. The new data appeared from the 2015/16 Preliminary Estimates.The change gives a more accurate classification of the level of fraud and claimant error across benefits, but it does mean that caution should be used in any comparisons between post 2014/15 results and earlier results. The change is thought to be the main reason for the increase in fraud and a corresponding fall in the level of claimant error, and it may have affected the overall level of total overpayments since its introduction after 2014/15. The new process has been applied to all of the continuously reviewed benefits but has had a particularly large effect on the Housing Benefit estimates. Fraud and Error in the Benefit System: FYE 2016 Estimates
Universal Credit estimates of fraud and error are included in the “Continuously Reviewed” estimates for the first time in the Preliminary 2015/16 results. Full reviews began in October 2014. There is an impact on the total overpayments and underpayments as previously fraud and error were estimated using a proxy measure. Fraud and Error in the Benefit System: FYE 2016 Estimates
Universal Credit introduced a methodology change where the statistics have been split into Reviewed and Cannot Review cases. The latter cases are included in the final statistics but calculated using assumptions as opposed to measured data. Fraud and Error in the Benefit System: FYE 2016 Estimates
The composition of the JSA sample has changed within this publication so that we no longer have a separate sample for newer cases. Fraud and Error in the Benefit System: FYE 2016 Estimates
JSA grossing no longer includes PSU or the yearly split in the calculation of the grossing factors; instead these have been calculated based solely on the client group. The grossing factors now mainly depend upon the national population for JSA, with the same grossing factor applied to all of the sample cases. This reduces the possibility of a relatively small number of cases with very high grossing factors within the sample having a large influence on the reported results, and has led to a decrease in the range of the confidence intervals around the central estimates for JSA. Fraud and Error in the Benefit System: FYE 2017 Estimates
PIP estimates for fraud and error were published for the first time within the 2016/17 final results. There is an impact on the total overpayments and underpayments as previously fraud and error on PIP was estimated using DLA as a proxy measure. Fraud and Error in the Benefit System: FYE 2017 Estimates
Rotational sampling has been introduced, for data collection between April 2017 and September 2019, so that a selection of hard to reach areas will only be selected within the sample at least once in a three-year period. This is the first set of published results that incorporate rotational sampling. Analysis completed on published statistics from previous years shows that the impact would be negligible – with changes to fraud and error levels of less than 0.05% for each benefit affected. Fraud and Error in the Benefit System: FYE 2018 Estimates
In the 2017/18 preliminary estimates we changed the way we gross ESA and PC estimates to using National grossing factors, bringing these benefits in line with JSA, as well as UC and PIP. National grossing tends to reduce the scope for individual cases to have a higher influence on the reported estimates and, consequently, leads to a decrease in the range of the confidence intervals . Fraud and Error in the Benefit System: FYE 2018 Estimates
In 2018, we started to simplify and align the methodology for each benefits reviewed that year. The 2017/18 final publication was the first one to use this new, standardised publication process for JSA, ESA, PC, SP (Official Error checks only), HB and UC. For HB, there was a significant change to how grossing and extrapolation were carried out, when bringing it in line with the other benefits. At an overall benefit view, the effect of the change is negligible (less than 0.1%) – however, when viewing at lower levels, such as the working age/pension age split, a bigger difference can be seen. The previous methodology increased the values associated with pension age claimants and decreased the values of the working age claimants more than the new process does. This also has a knock-on effect on published error reason categories at a total HB level, as reasons associated to a greater degree with either working age or pension age claimants will be affected. Fraud and Error in the Benefit System: FYE 2018 Estimates
The measurement processes are subject to a series of validation checks, which aim to check that the measurement methodology is being correctly implemented. A randomly selected sub-sample of cases are used to create an adjustment across the sample population, by assuming that the same rate of incorrectness/change applies to all cases. The movement to a new publication pipeline methodology and introduction of more refined and targeted data cleansing mean that this Data Quality Adjustment (DQA) is no longer required. The changes resulting from data cleansing and DQA checks are used as the final and correct outcome in the data processing and hence are incorporated directly into the calculations. Fraud and Error in the Benefit System: FYE 2018 Estimates
The 2017/18 final publication included UC estimates that were based on a composite measure of Live and Full Service cases. Calculation of these statistics required UC expenditure to be split by Full and Live Service so that the estimates for each service could be calculated separately before combining for the overall composite measure. The split applied was based on awards information from the UC caseloads data that underpins the national statistics. This was a temporary approach introduced for the 2017/18 final estimates, re-applied for 2018/19 estimates, and will not be required post 2018/19 as the estimates will be based on entirely Full Service samples. Fraud and Error in the Benefit System: FYE 2018 Estimates
A new assumption was introduced in the 2017/18 final publication for UC cases that did not have an effective review, primarily due to the claimant not engaging in the review process, resulting in their benefit claim being terminated. These cases are referred to as cannot review cases and are recorded as Fraud. The new assumption involves re-categorising these cases for reporting purposes following the outcome of checks to determine if the individuals had reclaimed benefit. The cases are re-categorised as Fraud, Not Fraud or inconclusive based on whether the individual reclaimed benefit or there was a suspicion of a specific type of fraud recorded on a case (for instance Capital Fraud). Inconclusive cases are not included in the headline statistics and are instead reported separately in a footnote in the publication. Fraud and Error in the Benefit System: FYE 2018 Estimates
For the 2017/18 final publication, changes were made to the Error Code Framework where we aligned the definition of ‘income – other’ and ‘income – other benefits’ across the six benefits reviewed this year, which has resulted in some small changes for Pension Credit and Housing Benefit pension age customers. Fraud and Error in the Benefit System: FYE 2018 Estimates
Following the evaluation of the pilot of desk based reviews, for the working age passported Housing Benefit client group, between April 2017 and September 2018, it has been decided to continue with face to face reviews for this client group; therefore, the 2017/18 final and the 2018/19 publications are based solely on face to face review cases. The consequence is that for these two publications the sample sizes were smaller for this client group, and the confidence intervals may be wider than in previous years. Fraud and Error in the Benefit System: FYE 2018 Estimates
The UC fraud and error estimates in the published national statistics were previously based on Live Service cases only which is the intermediary system in place to administer UC until the full online service is fully rolled out. MVFE estimates were based on the assumption that Full Service fraud and error rates would be similar to those being found in Live Service. The 2018/19 Universal Credit fraud and error estimates are based on Live and Full service cases. In total 1,998 Universal Credit cases were sampled, around 70% of these cases were Full service. The inclusion of Full and Live Service cases required expenditure assumptions to be updated to reflect the latest proportional service splits. The assumption to re-categorise cases that did not have an effective review, introduced in 2017/18, has also been applied to 4% of sample cases in 2018/19. Fraud and Error in the Benefit System: FYE 2019 Estimates
Between the 2017/18 publications in May and December 2018, we started to simplify and align the methodology for each benefits reviewed this year. This new publication pipeline contained Jobseeker’s Allowance, Employment and Support Allowance, Pension Credit, State Pension (Official Error checks only), Housing Benefit, Universal Credit, Personal Independence Payment. Fraud and Error in the Benefit System: FYE 2019 Estimates
Rotational sampling was introduced so that a selection of hard to reach areas are only selected within the sample, at least once in a three-year period. This rotational sampling was introduced for data collection between April 2017 and September 2019. The areas excluded this year are: October 2018 – September 2019: Cornwall, West Devon, South Hams, Teignbridge, Torridge, Highlands, Moray, and Argyll and Bute. Fraud and Error in the Benefit System: FYE 2020 Estimates
We have strengthened the process for how we deal with combination errors on Housing Benefit as new data has become available to us. Fraud and Error in the Benefit System: FYE 2020 Estimates
We have strengthened the process for how we deal with Housing Benefit cases that have a whole award error and an underpayment. Previously the underpayment would have been taken off the whole award within the netting and capping part of the process, as we would not know how much of the underpayment was valid. In the new process we look to see what the end award of Housing Benefit is after the review, to ensure we more accurately reflect what the loss to the department would have been. For example, if the overpayment would have removed entitlement to Housing Benefit altogether, then we would remove the underpayment. Fraud and Error in the Benefit System: FYE 2020 Estimates
Removal of Arrears Advance: This adjustment only affected Housing Benefit. This adjustment was applied to account for cases that were in arrears or advance as these cases could have more/less error at the time the payment relates to. Fraud and Error in the Benefit System: FYE 2020 Estimates
In 2019/20 we have rolled out the Cannot Review assumption from Universal Credit to the other benefits measured in 2019/20 for cases that did not have an effective review, primarily due to the claimant not engaging in the review process, resulting in their benefit claim being terminated. These cases are referred to as cannot review cases and are recorded as Fraud. The new assumption involves re-categorising these cases for reporting purposes following the outcome of checks to determine if the individuals had reclaimed benefit. The cases are re-categorised as Fraud, Not Fraud or inconclusive based on whether the individual reclaimed benefit or there was a suspicion of a specific type of fraud recorded on a case (for instance Capital Fraud). Inconclusive cases are not included in the headline statistics and are instead reported separately in a footnote in the publication. Previously, Official Error was not netted off from Fraud and Claimant Error on Employment and Support Allowance, Jobseeker’s Allowance (when it was last measured in 2018/19), Pension Credit and Housing Benefit, as they are reviewed in two different periods. When a review is carried out for these cases it is noted whether the Official Error continues into the review week. In over 99% of these cases the error was still there in period when the Fraud and Claimant Error was checked, and therefore netting it off would give a more accurate reflection of the true loss to the department. The note on whether the Official Error continued to the point when the Fraud and Claimant Error check is carried out is not on Housing Benefit. However, since the other benefits mentioned above and Housing Benefit have the same time lag between the Official Error checks and the Fraud/Claimant Error checks, we are confident we can apply this to Housing Benefit as well. Fraud and Error in the Benefit System: FYE 2020 Estimates
We have rolled out the Universal Credit approach for dealing with multiple whole award errors to the other measured benefits. Fraud and Error in the Benefit System: FYE 2020 Estimates
Removal of the Net Programme Value adjustment: This adjustment is only made on cases where a Living Together error was found. Previously we would have looked to see if they were still eligible for benefits after the review, and then changed the whole award Living Together error, to the difference between the amount of DWP benefits the claimant and partner were getting before the review and the amount of DWP benefits the claimant and partner were getting after the review. This was done to more accurately reflect the loss to the department. We have decided to remove this adjustment as it only affects a small number of cases (there were less than 10 of these cases in 2018/19). Fraud and Error in the Benefit System: FYE 2020 Estimates
Move to monthly grossing: We have rolled out the Universal Credit approach for grossing to the other benefits. Fraud and Error in the Benefit System: FYE 2020 Estimates
This year we have made a change to how we attribute the amount overpaid or underpaid to error reasons. For cases that have multiple errors, when capping the error values, we will attribute amounts to reasons in order of which we are most certain of. Any Fraud that is Causal Link (Low Suspicion) has been recategorised to a new category of “Failure to provide evidence/fully engage in the process”. This change has no effect on the amount overpaid or underpaid at a total level or an error type level (i.e. Fraud, Official Error, Claimant Error). Fraud and Error in the Benefit System: FYE 2021 Estimates
State Pension changes: Prior to this year we removed some accounting errors from SP, where an overpayment (or underpayment) error on SP was offset by an equivalent underpayment (or overpayment) of the same amount on Pension Credit. However, this was not the correct approach. We have now stopped removing these accounting errors from within SP this year. The impact of making this change is small (adding £1m to overpayments this year and £8m to underpayments). Fraud and Error in the Benefit System: FYE 2021 Estimates
State Pension changes: We have made a change to remove cases that have a deemed error on them from the calculation of the Official Error rate. Fraud and Error in the Benefit System: FYE 2021 Estimates
State Pension changes: We have made a small change to the calculation of our overall estimates for SP to better reflect changes in the split of expenditure each year between claimants resident in GB and those resident overseas. Fraud and Error in the Benefit System: FYE 2021 Estimates
State Pension changes: We have also made changes to remove the proportion of fraud and error relating to Dependency Increases from our SP estimates. This has only affected the estimates for FYE 2021 and not previous years. Fraud and Error in the Benefit System: FYE 2021 Estimates
The pandemic has driven a move to complete almost all reviews by telephone rather than face to face home visits. Fraud and Error in the Benefit System: FYE 2022 Estimates
Housing Benefit coverage: In FYE 2022, only non-passported working age claims were reviewed. This means that in FYE 2022, the non-passported working age estimates relate to reviews undertaken in FYE 2022, while the passported pension age estimates relate to reviews undertaken in FYE 2020, and the estimates for the remaining groups (passported working age and non-passported pension age) relate to reviews undertaken in FYE 2019. The rates of fraud and error found when each group was last reviewed were applied to the FYE 2022 expenditure, to calculate the total HB estimate for FYE 2022. Fraud and Error in the Benefit System: FYE 2022 Estimates
In order to calculate the monetary value of fraud and error and the proportion of expenditure overpaid we use DWP expenditure figures. Within these figures any case that was getting Housing Benefit and Universal Credit was classed as a non-passported Housing Benefit case. Although not strictly a passporting benefit, those getting Universal Credit and Housing Benefit are treated in a similar way to passported cases (if they are entitled to any Universal Credit their Housing Benefit is paid in full). Due to this we have made the change to the expenditure on Housing Benefit to classify these cases as passported and revised the last year’s figures. Fraud and Error in the Benefit System: FYE 2022 Estimates
Proportion of claims with fraud or an error: We introduced a “net” measure of the proportion of claims with fraud or an error to supplement existing tables. This measure is consistent with the calculation of the headline Monetary Value of Fraud and Error (MVFE) and is now reported alongside MVFE in the main statistical release. Fraud and Error in the Benefit System: FYE 2023 Estimates
De minimis: Following a review, the removal of all uprating errors of 10p or less on State Pension has been extended to all benefits and reasons for error. Updated figures using this methodology have also been produced for FYE 2022. Fraud and Error in the Benefit System: FYE 2023 Estimates
Housing Costs Capping on Universal Credit: The value of Housing Costs overpayments will now be limited (capped) to the amount of housing element in payment on a UC claim. This change has been made to more accurately reflect the amount of housing element that is overpaid on UC. Fraud and Error in the Benefit System: FYE 2023 Estimates
Reclassification of Failure to Provide Evidence Fraud: If a claimant fails to provide requested evidence after a Performance Measurement review, their benefit is suspended and then subsequently terminated. A whole award error is then recorded relating to the evidence that we failed to receive (e.g., if a bank statement which was requested to verify Capital was not received, then a whole award Capital Fraud is recorded). However, if we had no real suspicions about the evidence apart from the claimant not providing it, we reclassify the whole award error into Failure to provide evidence. Fraud and Error in the Benefit System: FYE 2023 Estimates
We have adopted a data driven approach to attempt to reclassify some of the failure to provide evidence errors into known reasons for Fraud. This involves looking at data we have on a case four months after classification, to see if this gives us a clearer picture of why they chose not to provide evidence. If this can be determined, we reclassify the error again from Failure to Provide into the error reason the data suggests. Fraud and Error in the Benefit System: FYE 2023 Estimates
Inclusion of an estimate of the fraud and error on Cost of Living Payments: Cost of Living Payments were made by the department for the first time in FYE 2023. These payments were to give claimants extra support during the current cost of living crisis and were paid in addition to any benefit that qualified a claimant for the payment. Since the statistics give an estimate of fraud and error on all benefit expenditure, we have included an estimate of the amount of fraud and error associated with these payments in our publication. Benefit reviews on Cost of Living Payments have not been carried out so an estimate has been derived. If a claimant is not eligible to receive benefit, then they would also not be eligible to receive a Cost of Living Payment. Therefore, to derive an estimate for the rate of fraud and error on these payments we have used the rate of cases that lose entitlement on the qualifying benefits (the majority of which have been measured in the current or recent years). Fraud and Error in the Benefit System: FYE 2023 Estimates
Changes to proxy benefits used for unreviewed benefits: For benefits which we have never measured we use a rate from a similar benefit we are currently measuring or have previously measured as a proxy. Alongside Cost of Living Payments being included for the first time we have carried out a full review of all the proxy benefits used in an effort to more accurately estimate the fraud and error on the unreviewed benefits. Fraud and Error in the Benefit System: FYE 2023 Estimates

ISBN: 978-1-78659-670-3