Copyright:
|
Disclaimer: The designations employed and the presentation of material throughout this publication do not imply the expression of any opinion whatsoever on the part of UNESCO concerning the legal status of any country, territory, city or area or of its authorities, or concerning the delimitation of its frontiers or boundaries. The ideas and opinions expressed in this publication are those of the authors; they are not necessarily those of UNESCO and do not commit the Organization. |
Citation: Symonds, E., Verbyla, M.E. and Mihelcic, J.M. (2019). Can farmers in Bolivia safely irrigate non-edible crops with treated wastewater? In: J.B. Rose and B. Jiménez-Cisneros, (eds) Water and Sanitation for the 21st Century: Health and Microbiological Aspects of Excreta and Wastewater Management (Global Water Pathogen Project). (S. Petterson and G. Medema (eds) Part 5 Case Studies), Michigan State University, E. Lansing, MI, UNESCO. https://doi.org/10.14321/waterpathogens.71 |
Last published: March 11, 2019 |
This case study aimed to determine if farmers, in low income countries, can safely reuse treated wastewater from an existing waste stabilization pond (WSP) system for irrigation, or are additional control measures or treatment processes required to reduce exposure to viral pathogens and meet a specified health target?
The study took place in a town, located in a culturally diverse region of the Caranavi province of Bolivia near the Alto Beni River, an important inland fishery system in the Amazon River basin. The local economy is driven by citrus fruit production for domestic sale and cacao beans for factories that manufacture chocolate. Many farmers chew coca leaves while working, resulting in frequent hand-to-mouth contact. Reclaimed wastewater can provide a local source of irrigation water that contains valuable nutrients and may be less carbon intensive than other sources. Like many areas of the world, most population growth will occur in small cities, such as the one studied here, that are closely linked to agricultural zones.
Figure 1. Community-operated waste stabilization pond (WSP) system with (a) a facultative pond and (b-c) two maturation ponds in series (left); case study site location (right; photo by M.E. Verbyla).
The wastewater treatment system serves 780 people and consists of flush toilets, a gravity-driven conveyance network, and three WSP in series. While it provided high removal of faecal coliforms, limited virus removal was measured. Treated effluent is discharged to a nearby surface water, but some farmers would like to use the effluent for irrigation. This sanitation system is managed and operated by a volunteer community water committee.
Minor additional control measures are needed to reduce the risk of virus exposure during farming and meet the specified health target for this study. It is better to use at least two of these measures in combination to create “multiple barriers” for pathogen control. If one barrier fails, others will still provide some protection.
Read more?
scroll down for a more detailed case study description
Wastewater use in agriculture facilitates water and nutrient recovery, offsetting energy needs for food production and reducing the degradation of aquatic ecosystems (Hamilton et al., 2007). Currently, 20 million hectares of land are irrigated with wastewater (Raschid-Sally and Jayakody, 2008). The extent of wastewater irrigation will likely increase in the future because of water scarcity, population growth, and the adoption of the Sustainable Development Goals (SDGs), which include a target to increase water recycling and safe reuse globally. Reclaiming treated wastewater is also beneficial because it applies nitrogen and phosphorus to land instead of surface water, which reduces the eutrophication potential of the sanitation system. Reusing treated wastewater may also lower the carbon footprint and embodied energy of sanitation systems, especially systems with high material and energy inputs (Cornejo et al., 2013). The World Health Organization (WHO) recommends a systematic risk-based approach to assess wastewater reuse via Sanitation Safety Planning (SSP; WHO, 2016), with a maximum health burden of 10-6 disability-adjusted life years (DALYs) lost per person per year. Since it has been suggested that 10-4 DALYs may be a more appropriate initial target for regions with high diarrheal disease burdens (Mara et al., 2010), the target of 10-4 DALYs was selected to evaluate the risk of reusing water from a three-pond waste stabilization pond (WSP) system in Bolivia.
While there are many ways to reduce pathogen concentrations in wastewater prior to reuse, WSPs are extremely prevalent worldwide and facilitate natural disinfection and removal processes without requiring high energy or material inputs (Kumar and Asolekar, 2016; Maynard et al., 1999; Oakley, 2005; Verbyla and Mihelcic, 2015; Verbyla et al., 2013a). Pathogen reduction is primarily achieved in tertiary maturation or polishing ponds. Based on Verbyla et al. ,2013a), this system provided an average 3.4-log10 removal of faecal coliforms. Since enteric viruses are often more resistant to treatment, enteric virus reference pathogens were directly measured. This case study highlights a quantitative microbial risk assessment (QMRA) of agricultural irrigation with treated effluent from a community-managed wastewater treatment system in Bolivia consisting of three WSPs in series (Figure 1; Symonds et al., 2014). The QMRA determines the additional log10 enteric virus reductions required to safely reuse the treated effluent and considers the health risks to adult farmers as well as children at play in irrigation fields. The setting is like many areas of the world, where most population growth will occur in small cities closely linked to agricultural zones (Verbyla et al., 2013a).
The purpose of the QMRA was to determine the additional log10 enteric virus reductions necessary to ensure the safe reuse of effluent from a three-pond community-managed wastewater treatment systems for irrigation. The work is based on a previously published study (Symonds et al., 2014).
The scope was defined by:
Hazard identification: Enteric viruses, represented by norovirus (measured by RT-qPCR) for adult farmers and rotavirus (measured by RT-qPCR) for children <5 years.
Exposure pathways: two exposure pathways were considered:
Health outcome: DALYs lost per person per year was selected as the health outcome, with a target of 10-4 DALYs per person, since Bolivia has a high diarrheal disease burden (Mara et al., 2010).
Source: The concentrations of norovirus and rotavirus were determined by molecular methods (RT-qPCR) from composite samples of treated wastewater collected over a 24-hour period in June 2012. Since this study used molecular methods to determine rotavirus concentrations and culture-based methods were used to develop the dose-response relationship (Ward et al. 1986), it was necessary to harmonize rotavirus concentrations using a ratio 1:1000 to 1:1900 gene copies to focus-forming units (Mok and Hamilton, 2014). Such an adjustment was not needed for norovirus due to congruent methods used in this study and in the dose-response studies.
Barriers/controls: The risk of enteric virus illness from wastewater reuse for a three-pond wastewater treatment system was executed with respect to farmers and children playing in fields irrigated with treated effluent (Symonds et al., 2014).
Exposure :The assumed amount of virus ingested during exposure to treated wastewater effluent was determined based upon the assumed volume of effluent ingested and the concentration of enteric viruses in the effluent. It was assumed that adult famers and children playing in fields ingested the equivalent of 1.0 mL of wastewater effluent per day (Ottoson and Stenström, 2003), during 75 days/year for farmers and 150 days/year for children (Mara et al., 2007; Seidu et al., 2008). Log-normal distributions of virus concentrations were assumed, based on those measured in the treated wastewater effluents (Table 1).
Dose-response models were used to determine the additional virus removal necessary to safely reuse of the wastewater treatment system effluent with respect to farmers and children in fields. The hypergeometric model (Teunis et al., 2008) with a Pfaff transformation (Barker et al., 2013; Mok et al., 2014) was used for norovirus, where the probability of infection was calculated as:
$$P\scriptsize inf\tiny NV = \normalsize1-(\scriptsize2\normalsize F \scriptsize 1 \normalsize(\beta\tiny NV,\frac{\normalsize C \tiny NV \normalsize V(1-a\tiny NV \normalsize)}{\normalsize a \tiny NV},\normalsize \alpha \tiny NV\normalsize-\beta\tiny NV\normalsize;a \tiny NV\normalsize )(\frac{1}{1-a \tiny NV \normalsize})^\frac{-(\normalsize C \tiny NV \normalsize V(1-a \tiny NV \normalsize)}{\normalsize a \tiny NV})$$ (1)
where αNV=0.04; βNV=0.055; aNV=0.9997 (Teunis et al., 2008); and where cNV is the concentration of norovirus and V is the volume of water ingested. Not everyone who becomes infected develops an illness (there is the possibility that some become ‘silent carriers’); therefore, a conditional probability of norovirus illness (the proportion of infected individuals developing symptoms of an illness) was calculated using:
$$P_{ill\inf_{NV}}=1-(1-\eta_{NV}c_{NV}V)^{-r_{NV}} $$ where $\eta_{NV}=0.00255; r_{NV}=0.086$ (2)
Rotavirus probability of infection was calculated using the exact beta-Poisson model (Teunis and Havelaar, 2000):
$${p_{inf}}_{RV}=1-_{1}F_{1}(\alpha_{RV},\alpha_{RV}+\beta_{RV},-c_{RV}V)$$ where $\alpha_{RV}=0.167;\beta_{RV}=0.191$ (3)
and the conditional probability of rotavirus illness given infection was determined assuming a simple ratio of 0.9 (Havelaar and Melse, 2003):
$$p_{ill\mid inf_{RV} }=p_{inf_{RV}}\cdot0.9$$ (4)
The probability of contracting an illness that would cause some type of disease burden was calculated as:
$$p_{ill}=p_{inf}\cdot p_{ill\mid inf}$$ (5)
To normalize the probability of illness per year for the two groups exposed for a different number of days per year, the following equation was used, where n is the number of days per year of exposure:
$$p_{ill_{annual}}=1-(1-p_{ill_{daily}})^n$$ (6)
It is important to consider the local context of exposure when completing QMRAs, especially when locally-derived exposure data is not available. This can be done using a sensitivity analysis. For this study, it was assumed that farmers accidentally ingest 1.0 mL of irrigation water per day while working. However, this assumption came from a publication written within the context of irrigation practices in Sweden (Ottoson and Stenström, 2003). In Bolivia, some farmers chew coca leaves while working, a practice that implies frequent hand-to-mouth contact and creates the possibility that greater volumes of irrigation water and/or soil are accidentally ingested. A sensitivity analysis revealed that if the amount of water accidentally ingested were doubled (increased from 1.0 mL to 2.0 mL per day), an additional virus reduction of 0.3-log10 units (in addition to the log10 reductions presented in Figure 2) would be required.
The reuse of the effluent from both wastewater treatment systems for restricted agricultural irrigation exceeded the health benchmark of 10-4 DALYs for adult farmers and children. Based upon a conservative interpretation of the QMRA (the upper 97.5% confidence interval), an additional 5.2-log10 rotavirus reduction would be required to ensure the safety of young children playing in irrigation fields. To ensure the safety of farmers irrigating with treated effluent, up to 1.6-log10 of additional norovirus reduction would be required.
Therefore, the following interventions are recommended: