Votre recherche
Résultats 72 ressources
-
RÉSUMÉ: Les inondations sont considérées comme l'un des risques naturels les plus dangereux au monde. Plusieurs pays souffrent des conséquences néfastes des inondations. Au Canada, plusieurs provinces ont subi des inondations au cours du siècle dernier. Par exemple, la rivière des Outaouais a été confrontée à de nombreuses inondations comme en 2017 et 2019. La population d'Ottawa continue à augmenter d'une année à l'autre. C'est pour cela que nous avons choisi la rivière des Outaouais comme étude de cas pour ce projet dans le but de protéger la société contre les risques causés par les inondations. Les pays adoptent plusieurs solutions basées sur différentes méthodes afin de minimiser les dommages causés par les crues. La plupart des scientifiques s'accordent que la prévision des crues est la meilleure façon de limiter les conséquences des crues. Les systèmes de prévision des crues sont indispensables dans les pays fréquemment confrontés à des crues. Ils visent à fournir un long délai d'exécution et à fournir aux autorités et aux décideurs des informations suffisantes. Par conséquent, ils auront suffisamment de temps pour prendre les mesures adéquates pour sauver la vie de la population et limiter les catastrophes économiques dues aux inondations. ABSTRACT: Floods are one of the most catastrophic natural disasters in Canada and around the world that can cause loss of life and damages to properties and infrastructures. Saguenay flood (1996), southern Alberta flood (2013), and Ottawa floods (2017, 2019), are a few examples of Canadian floods with tremendous socio-economic impacts. Flood forecasting and predicting its characteristics (e.g., its magnitude and extent) has an important role in preventing and mitigating such flood impacts. Particularly, short-term forecasting is crucial for early warning systems and emergency response to floods. This study presents an integrated hydraulic-hydrologic modeling system for flood prediction. In this system, the Delft3D two-dimensional hydrodynamic model is connected with a HEC-HMS hydrologic model and observation data to provide an automatic exchange of data and results. Delft3D and HEC-HMS were chosen for this study because they were widely used and provided good results. In addition, they were applied in several flood forecasting studies. The prediction weather data and watershed characteristics provide input to the hydrological model to predict streamflow conditions, which are then automatically fed into the hydrodynamic model. The hydrodynamic model simulates the flood characteristics such as water level, 2D depth-averaged velocity field, and flood extent.
-
The frequency of natural hazards in North America presents a significant challenge for governments due to the damages they cause to the environment. Floods are severe hydrological events caused by spring snowmelt and intense rain events. Flood frequency analysis studies assumes that annual peak flood events occur independently of each other, regardless of previous flood events (the independent and identically distributed (i.i.d.) assumption); however, annual peak flood records do not necessarily appear to conform to these assumptions. First, a review of the literature on the effects of climate oscillations on extreme flood frequencies in North America was conducted. Then, the i.i.d. flood event assumption was tested by analyzing the effects of the Pacific Decadal Oscillation (PDO) and Atlantic Multidecadal Oscillation (AMO) on 250 naturally flowing annual peak flood records across the entire western North American margin. Using permutation tests on quantile-quantile (Q-Q) plots, I found that the PDO has a greater impact on the magnitude of annual peak floods than the AMO. Twenty-six percent of the gauges have higher magnitude annual floods depending on the PDO phase (p < 0.1). Next, I examined the interacting effects of the PDO and AMO on the frequencies of lower and upper quartile annual peak floods, and found reinforcing, cancelling, and dominating effects. Since these two climate oscillations have significant effects on the magnitudes of annual peak floods, the i.i.d. assumption does not hold. Hence, I advocate for the need to re-assess baseline flood analysis in western North America to improve flood management strategies.
-
Abstract Global warming is causing glaciers in the Caucasus Mountains and around the world to lose mass at an accelerated pace. As a result of this rapid retreat, significant parts of the glacierized surface area can be covered with debris deposits, often making them indistinguishable from the surrounding land surface by optical remote-sensing systems. Here, we present the DebCovG-carto toolbox to delineate debris-covered and debris-free glacier surfaces from non-glacierized regions. The algorithm uses synthetic aperture radar-derived coherence images and the normalized difference snow index applied to optical satellite data. Validating the remotely-sensed boundaries of Ushba and Chalaati glaciers using field GPS data demonstrates that the use of pairs of Sentinel-1 images (2019) from identical ascending and descending orbits can substantially improve debris-covered glacier surface detection. The DebCovG-carto toolbox leverages multiple orbits to automate the mapping of debris-covered glacier surfaces. This new automatic method offers the possibility of quickly correcting glacier mapping errors caused by the presence of debris and makes automatic mapping of glacierized surfaces considerably faster than the use of other subjective methods.
-
Abstract Temporal variations in concentrations of pathogenic microorganisms in surface waters are well known to be influenced by hydrometeorological events. Reasonable methods for accounting for microbial peaks in the quantification of drinking water treatment requirements need to be addressed. Here, we applied a novel method for data collection and model validation to explicitly account for weather events (rainfall, snowmelt) when concentrations of pathogens are estimated in source water. Online in situ β ‐ d ‐glucuronidase activity measurements were used to trigger sequential grab sampling of source water to quantify Cryptosporidium and Giardia concentrations during rainfall and snowmelt events at an urban and an agricultural drinking water treatment plant in Quebec, Canada. We then evaluate if mixed Poisson distributions fitted to monthly sampling data ( = 30 samples) could accurately predict daily mean concentrations during these events. We found that using the gamma distribution underestimated high Cryptosporidium and Giardia concentrations measured with routine or event‐based monitoring. However, the log‐normal distribution accurately predicted these high concentrations. The selection of a log‐normal distribution in preference to a gamma distribution increased the annual mean concentration by less than 0.1‐log but increased the upper bound of the 95% credibility interval on the annual mean by about 0.5‐log. Therefore, considering parametric uncertainty in an exposure assessment is essential to account for microbial peaks in risk assessment.
-
Conventional processes (coagulation, flocculation, sedimentation, and filtration) are widely used in drinking water treatment plants and are considered a good treatment strategy to eliminate cyanobacterial cells and cell-bound cyanotoxins. The diversity of cyanobacteria was investigated using taxonomic cell counts and shotgun metagenomics over two seasons in a drinking water treatment plant before, during, and after the bloom. Changes in the community structure over time at the phylum, genus, and species levels were monitored in samples retrieved from raw water (RW), sludge in the holding tank (ST), and sludge supernatant (SST). Aphanothece clathrata brevis, Microcystis aeruginosa, Dolichospermum spiroides, and Chroococcus minimus were predominant species detected in RW by taxonomic cell counts. Shotgun metagenomics revealed that Proteobacteria was the predominant phylum in RW before and after the cyanobacterial bloom. Taxonomic cell counts and shotgun metagenomic showed that the Dolichospermum bloom occurred inside the plant. Cyanobacteria and Bacteroidetes were the major bacterial phyla during the bloom. Shotgun metagenomics also showed that Synechococcus, Microcystis, and Dolichospermum were the predominant detected cyanobacterial genera in the samples. Conventional treatment removed more than 92% of cyanobacterial cells but led to cell accumulation in the sludge up to 31 times more than in the RW influx. Coagulation/sedimentation selectively removed more than 96% of Microcystis and Dolichospermum. Cyanobacterial community in the sludge varied from raw water to sludge during sludge storage (1–13 days). This variation was due to the selective removal of coagulation/sedimentation as well as the accumulation of captured cells over the period of storage time. However, the prediction of the cyanobacterial community composition in the SST remained a challenge. Among nutrient parameters, orthophosphate availability was related to community profile in RW samples, whereas communities in ST were influenced by total nitrogen, Kjeldahl nitrogen (N- Kjeldahl), total and particulate phosphorous, and total organic carbon (TOC). No trend was observed on the impact of nutrients on SST communities. This study profiled new health-related, environmental, and technical challenges for the production of drinking water due to the complex fate of cyanobacteria in cyanobacteria-laden sludge and supernatant.
-
Waterborne pathogens are heterogeneously distributed across various spatiotemporal scales in water resources, and representative sampling is therefore crucial for accurate risk assessment. Since regulatory monitoring of microbiological water quality is usually conducted at fixed time intervals, it can miss short-term fecal contamination episodes and underestimate underlying microbial risks. In the present paper, we developed a new automated sampling methodology based on near real-time measurement of a biochemical indicator of fecal pollution. Online monitoring of β-D-glucuronidase (GLUC) activity was used to trigger an automated sampler during fecal contamination events in a drinking water supply and at an urban beach. Significant increases in protozoan parasites, microbial source tracking markers and E. coli were measured during short-term (<24 h) fecal pollution episodes, emphasizing the intermittent nature of their occurrence in water. Synchronous triggering of the automated sampler with online GLUC activity measurements further revealed a tight association between the biochemical indicator and culturable E. coli. The proposed event sampling methodology is versatile and in addition to the two triggering modes validated here, others can be designed based on specific needs and local settings. In support to regulatory monitoring schemes, it should ultimately help gathering crucial data on waterborne pathogens more efficiently during episodic fecal pollution events.
-
Abstract When the Syrian war erupted in 2011, the Lebanese government withdrew from managing the influx of Syrian refugees. Three years later, Lebanon’s Council of Ministers set new regulations for Syrians with the purpose of reducing access to territory and persuading refugees to leave the country. This article analyses the reasons for and the outcomes of Lebanon’s response to the refugee crisis before and after 2014. It then examines, through a qualitative exploratory approach and based on two longitudinal case studies, the impact of Lebanese regulations. In both cases, the so-called ‘temporary gatherings’ became permanent settlements beyond the government’s control. The government’s strategy backfired: in attempting to avoid ghettos, it created them. We conclude that when refugee situations become protracted, most efforts aimed at excluding refugees fail. Excluding refugees increases their vulnerability and reduces their chances of repatriation or resettlement. To prevent this, we argue that hosting policies must lead to the temporary integration of refugees within urban systems and public institutions.
-
The numerical modeling of sediment transport under wave impact is challenging because of the complex nature of the triple wave–structure–sediment interaction. This study presents three-dimensional numerical modeling of sediment scouring due to non-breaking wave impact on a vertical seawall. The Navier–Stokes–Exner equations are approximated to calculate the full evolution of flow fields and morphodynamic responses. The bed erosion model is based on the van Rijn formulation with a mass-conservative sand-slide algorithm. The numerical solution is obtained by using a projection method and a fully implicit second-order unstructured finite-volume method in a σ-coordinate computational domain. This coordinate system is employed to accurately represent the free-surface elevation and sediment/water interface evolution. Experimental results of the velocity field, surface wave motion, and scour hole formation hole are used to compare and demonstrate the proposed numerical method’s capabilities to model the seawall scour.
-
Several statistical methods were used to analyze the spatio-temporal variability of daily minimum extreme flows (DMEF) in 17 watersheds—divided into three homogenous hydroclimatic regions of southern Quebec—during the transitional seasons (spring and fall), during the 1930–2019 period. Regarding spatial variability, there was a clear difference between the south and north shores of the St. Lawrence River, south of 47° N. DMEF were lower in the more agricultural watersheds on the south shore during transitional seasons compared to those on the north shore. A correlation analysis showed that this difference in flows was mainly due to more agricultural areas ((larger area (>20%) on the south than on the north shore (<5%)). An analysis of the long-term trend of these flows showed that the DMEF of south-shore rivers have increased significantly since the 1960s, during the fall (October to December), due to an increase in rainfall and a reduction in cultivated land, which increased the infiltration in the region. Although there was little difference between the two shores in the spring (April to June), we observed a decrease in minimum extreme flows in half (50%) of the south-shore rivers located north of 47° N.
-
Abstract. Climate change impact studies require a reference climatological dataset providing a baseline period to assess future changes and post-process climate model biases. High-resolution gridded precipitation and temperature datasets interpolated from weather stations are available in regions of high-density networks of weather stations, as is the case in most parts of Europe and the United States. In many of the world's regions, however, the low density of observational networks renders gauge-based datasets highly uncertain. Satellite, reanalysis and merged product datasets have been used to overcome this deficiency. However, it is not known how much uncertainty the choice of a reference dataset may bring to impact studies. To tackle this issue, this study compares nine precipitation and two temperature datasets over 1145 African catchments to evaluate the dataset uncertainty contribution to the results of climate change studies. These deterministic datasets all cover a common 30-year period needed to define the reference period climate. The precipitation datasets include two gauge-only products (GPCC and CPC Unified), two satellite products (CHIRPS and PERSIANN-CDR) corrected using ground-based observations, four reanalysis products (JRA55, NCEP-CFSR, ERA-I and ERA5) and one merged gauged, satellite and reanalysis product (MSWEP). The temperature datasets include one gauged-only (CPC Unified) product and one reanalysis (ERA5) product. All combinations of these precipitation and temperature datasets were used to assess changes in future streamflows. To assess dataset uncertainty against that of other sources of uncertainty, the climate change impact study used a top-down hydroclimatic modeling chain using 10 CMIP5 (fifth Coupled Model Intercomparison Project) general circulation models (GCMs) under RCP8.5 and two lumped hydrological models (HMETS and GR4J) to generate future streamflows over the 2071–2100 period. Variance decomposition was performed to compare how much the different uncertainty sources contribute to actual uncertainty. Results show that all precipitation and temperature datasets provide good streamflow simulations over the reference period, but four precipitation datasets outperformed the others for most catchments. They are, in order, MSWEP, CHIRPS, PERSIANN and ERA5. For the present study, the two-member ensemble of temperature datasets provided negligible levels of uncertainty. However, the ensemble of nine precipitation datasets provided uncertainty that was equal to or larger than that related to GCMs for most of the streamflow metrics and over most of the catchments. A selection of the four best-performing reference datasets (credibility ensemble) significantly reduced the uncertainty attributed to precipitation for most metrics but still remained the main source of uncertainty for some streamflow metrics. The choice of a reference dataset can therefore be critical to climate change impact studies as apparently small differences between datasets over a common reference period can propagate to generate large amounts of uncertainty in future climate streamflows.