Food fortification for impact: a data-driven approach

2016 
The importance of scientific evidence to guide nutrition policy and programme design is well established.1 Nevertheless, there are still critical gaps in the evidence to inform nutrition programme priorities and achievements,2 and strong pleas have been made for improving the collection and use of evidence in the nutrition sector.3 Nutrition interventions are founded on a strong evidence base from clinical trials, the systematic review of those trials and the translation of this evidence into global guidance.1 However, we have far less evidence, particularly for functional outcomes,4 on the impact of nutrition programmes that incorporate – but often go beyond – single interventions. We also know less about the pathways by which impacts are achieved in a programmatic context.5 Despite the importance of programme coverage for achieving impact, only a few nutrition interventions have sound coverage estimates, including high-dose vitamin A supplementation, household use of iodized salt and, in some countries, community management of acute malnutrition.2 In an effort to address these evidence gaps, the Global Nutrition Report of 20142 and the Micronutrient Forum 2014 Global Conference proceedings6 specifically highlighted the need to pay more attention to programme coverage as the main approach to assessing the availability, access and utilization of nutrition programmes. Moreover, the World Health Organization, as part of the Comprehensive Implementation Plan on Maternal, Infant and Young Child Nutrition,7 has called for renewed and coordinated efforts to better track programme coverage and to build systems for data-driven decision-making. Fortification of staple foods is usually mandated and regulated by governments and, when this functions well, is enforced at the point of production or import. But when foods are then sold on the market, it is difficult for health information or surveillance systems to track their use. Fortification of staple foods by the addition of vitamins and/or minerals to food at processing is widely recognized to be a cost–effective public health intervention that can reach large segments of the population.8 For example, goitre and other severe iodine deficiency disorders have been virtually eliminated in most countries around the world due to iodization of salt,9 and the contribution of wheat flour fortification with folic acid to reductions in the incidence of neural tube defects10 in several countries is well documented. At the same time there is still some controversy around the safety of long-term high intakes of some nutrients, most notably folic acid,11 thus again highlighting the need to measure dietary sources of nutrients and to track additional intakes due to the consumption of fortified foods. While in most countries governments regulate food fortification – i.e. what can be fortified, with which nutrients and at what levels – monitoring of compliance and enforcement of regulations needs to be strengthened considerably in many countries.12 Further compounding the issue are the lack of data on dietary intakes in most countries and how dietary patterns vary by population sub-groups – e.g. by socioeconomic status, geographical region or ethnicity. Such information is important to identify the dietary nutrient gaps that need to be addressed, as well as which nutrition strategies may be viable from a programmatic perspective. Data about intakes are required to set appropriate levels of fortificants, as well as to assess the extent to which those in need are actually consuming the foods, with what frequency and quantity.8 Some efforts have been made to fill these gaps, via analysis of household income and expenditure data at the national level,13 but the lack of detail about the types of food consumed and individuals’ consumption levels may limit the utility of these data to fully inform food fortification programmes.14 The Fortification Rapid Assessment Tool (FRAT)15 and more recently the Fortification Assessment Coverage Tool (FACT)16 were designed to provide more comprehensive information on consumption of fortified foods and foods that could potentially be fortified. Indicators of the use of fortified foods, and even food sample collection (e.g. salt samples) to verify compliance, have been incorporated into national surveillance systems in a few countries – e.g. Nicaragua17 – and modules from FRAT or FACT could potentially be adapted and included as well. The big question remains: how to garner the resources and political commitment needed for the generation and use of evidence for programme decision-making in nutrition broadly and food fortification specifically? This question was high on the agenda at the first Global Summit on Food Fortification in Arusha, United Republic of Tanzania from 9 to 11 September 2015. Government delegations from 29 countries from Africa, Asia and Central and South America, and experts and representatives from donors, United Nations agencies, nongovernmental organizations working in food fortification and the private sector, met to discuss and debate the state of food fortification in their countries and globally. The Summit declaration18 highlighted the need for increased investment and use of evidence to inform fortification programmes as part of national nutrition strategies. We hope that one of the lasting outcomes of this meeting will be greater commitment from countries and donors for evidence-driven decision-making on food fortification, as part of micronutrient deficiency control strategies – where it is needed and where it can have impact.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    11
    References
    7
    Citations
    NaN
    KQI
    []