There's a hole in the data -Kiran Bhatty & Dipa Sinha
-The Indian Express
The state has failed to create capacities for a timely, reliable, decentralised data regime.
The credibility of India’s data systems is under serious threat with the recent controversy over the employment data of the National Sample Survey. While the Census of India and the National Sample Survey Organisation (NSSO) have a good reputation, when it comes to data related to the social sector — health, education, nutrition — the situation, even with these sources (along with other large data sets), has been deficient on numerous counts.
One, the information collected is not available in real time or even annually. The NSSO collects data through specific rounds (health expenditure, debt etc.) which don’t have a fixed cycle unlike the consumption expenditure surveys, and the Census collects data once in 10 years. Budget allocations follow an annual cycle and policy pronouncements are not dovetailed to the years for which data is available. This raises important questions about the basis on which policies and plans are made. In the case of malnutrition, which is a problem needing urgent solutions, there was no independent data telling us what the trends are for a long time — the National Family Health Survey (NFHS-4) report came out in 2017 after a gap of over 10 years (NFHS-3 was in 2006). In the interim, major initiatives were planned for the eradication of malnutrition without any inkling of the situation on the ground or how it was changing. Similarly, data on learning levels was not collected consistently by the government, till 2017, and it is not known when the next round will be held or how long it will take for the data to be made available.
Two, there are inconsistencies in definitions and sampling frames across data sources and across time in the same data source. For instance, questions posed by the NSS for obtaining information on out-of-school children vary dramatically from those posed by the Census. As a result, the two arrive at vastly different numbers. Similarly, in the case of malnutrition data, there have been changes in the definitions used by NFHS across different rounds that make comparisons over time difficult. Periodicity of data collection also varies across sources, furthering difficulty in validation. Data validation plays an important part in improving the quality of data collected and ensuring authenticity, without which departments are basically shooting in the dark.
Three, the data collected in these surveys, is not geared towards policy or planning. The education rounds of NSS are part of the survey on social consumption, which in turn is for the purpose of making an assessment of the benefits derived by various sections of society from public expenditure incurred by the government. It provides no information on how the education system is functioning. As a result, several important indicators that would be of interest for planning or to the people, do not even figure in them. For instance, the different categories of teachers or their salaries is a not a data point in any data-set on education.
In the absence of regular large-scale survey data, what is available is the registry data collected by departments and ministries for monitoring of programmes. Unfortunately, these too suffer from gaps in information and are rarely used for programmatic purposes. At most, they are part of an accounting exercise. For instance, school surveys by the MHRD collect information on broad indicators of infrastructure and teacher availability (only two categories, whereas multiple exist) and student enrolment (but not attendance) and distribution of incentives. These take stock of the provisioning in schools, showcasing administrative efforts, but not functioning of the education system or real changes within it.
Please click here to read more.