An exploratory study on soil contamination of heavy metals was carried out surrounding Huludao zinc smelter in Liaoning province,
China. The distribution of total heavy metals and their chemical speciations were investigated. The correlations between heavy
metal speciations and soil pH values in corresponding sites were also analyzed. In general, Cd, Zn, Pb, Cu and As presented
a significant contamination in the area near the smelter, comparied with Environmental Quality Standards for Soils in China.
The geoaccumulation index showed the degree of contamination: Cd > Zn > Pb > Cu > As. There was no obvious pollution of Cr
and Ni in the studied area. The speciation analysis showed that the dominant fraction of Cd and Zn was the acid soluble fraction,
and the second was the residual fraction. Pb was mostly associated with the residual fraction, which constituted more than
50% of total concentration in all samples. Cu in residual fraction accounted for a high percentage (40–80%) of total concentration,
and the proportion of Cu in the oxidizable fraction is higher than that of other metals. The distribution pattern of Pb and
Zn was obviously affected by soil pH. It seemed that Pb and Zn content in acid solution fraction increased with increasing
soil pH values, while Cd content in acid soluble fraction accounted for more proportion in neutral and alkaline groups than
acidic one. The fraction distribution patterns of Cu in three pH groups were very similar and independent of soil pH values.
And the residual fraction of Cu took a predominant part (50%) of the total content. 相似文献
The Cu–Co–Ni Texeo mine has been the most important source of Cu in NW Spain since Roman times and now, approximately 40,000 m3 of wastes from mine and metallurgical operations, containing average concentrations of 9,263 mg kg−1 Cu, 1,100 mg kg−1 As, 549 mg kg−1 Co, and 840 mg kg−1 Ni, remain on-site. Since the cessation of the activity, the abandoned works, facilities and waste piles have been posing
a threat to the environment, derived from the release of toxic elements. In order to assess the potential environmental pollution
caused by the mining operations, a sequential sampling strategy was undertaken in wastes, soil, surface and groundwater, and
sediments. First, screening field tools were used to identify hotspots, before defining formal sampling strategies; so, in
the areas where anomalies were detected in a first sampling stage, a second detailed sampling campaign was undertaken. Metal
concentrations in the soils are highly above the local background, reaching up to 9,921 mg kg−1 Cu, 1,373 mg kg−1 As, 685 mg kg−1 Co, and 1,040 mg kg−1 Ni, among others. Copper concentrations downstream of the mine works reach values up to 1,869 μg l−1 and 240 mg kg−1 in surface water and stream sediments, respectively. Computer-based risk assessment for the site gives a carcinogenic risk
associated with the presence of As in surface waters and soils, and a health risk for long exposures; so, trigger levels of
these elements are high enough to warrant further investigation. 相似文献
Many different runout prediction methods can be applied to estimate the mobility of future debris flows during hazard assessment. The present article reviews the empirical, analytical, simple flow routing and numerical techniques. All these techniques were applied to back-calculate a debris flow, which occurred in 1982 at La Guingueta catchment, in the Eastern Pyrenees. A sensitivity analysis of input parameters was carried out, while special attention was paid to the influence of rheological parameters. We used the Voellmy fluid rheology for our analytical and numerical modelling, since this flow resistance law coincided best with field observations. The simulation results indicated that the “basal” friction coefficients rather affect the runout distance, while the “turbulence” terms mainly influence flow velocity. A comparison of the velocity computed on the fan showed that the analytical model calculated values similar to the numerical ones. The values of our rheological parameters calibrated at La Guingueta agree with data back-calculated for other debris flows. Empirical relationships represent another method to estimate total runout distance. The results confirmed that they contain an important uncertainty and they are strictly valid only for the conditions, which were the basis for their development. With regards to the simple flow routing algorithm, this methods could satisfactorily simulate the total area affected by the 1982 debris flow, but it was not able to directly calculate total runout distance and velocity. Finally, a suggestion on how different runout prediction methods can be applied to generate debris-flow hazard maps is presented. Taking into account the definition of hazard and intensity, the best choice would be to divide the resulting hazard maps into two types: “final hazard maps” and “preliminary hazard maps”. Only the use of numerical models provided final hazard maps, because they could incorporate different event magnitudes and they supplied output-values for intensity calculation. In contrast, empirical relationships and flow routing algorithms, or a combination of both, could be applied to create preliminary hazard maps. The present study only focussed on runout prediction methods. Other necessary tasks to complete the hazard assessment can be looked up in the “Guidelines for landslide susceptibility, hazard and risk zoning” included in this Special Issue. 相似文献
This paper describes a new procedure for assessing the ratio between in situ stresses in rock masses by means of K (K = σH / σv, being σH and σv principal stress) and tectonics for purposes of engineering geology and rock mechanics. The method combines the use of the logic decision tree and the empirical relationship between the Tectonic Stress Index, TSI, and a series of K in situ values obtained from an extensive database. The decision tree considers geological and geophysical factors affecting stress magnitudes both on the regional and local scale. The TSI index is defined by geological and geomechanical parameters. The method proposed provides an assessment of the magnitude of horizontal stresses of tectonic origin. Results for several regions of Europe are presented and the possible applications of the procedure are discussed. 相似文献
The paper is dedicated to the review of methods of seismic hazard analysis currently in use, analyzing the strengths and weaknesses of different approaches. The review is performed from the perspective of a user of the results of seismic hazard analysis for different applications such as the design of critical and general (non-critical) civil infrastructures, technical and financial risk analysis. A set of criteria is developed for and applied to an objective assessment of the capabilities of different analysis methods. It is demonstrated that traditional probabilistic seismic hazard analysis (PSHA) methods have significant deficiencies, thus limiting their practical applications. These deficiencies have their roots in the use of inadequate probabilistic models and insufficient understanding of modern concepts of risk analysis, as have been revealed in some recent large scale studies. These deficiencies result in the lack of ability of a correct treatment of dependencies between physical parameters and finally, in an incorrect treatment of uncertainties. As a consequence, results of PSHA studies have been found to be unrealistic in comparison with empirical information from the real world. The attempt to compensate these problems by a systematic use of expert elicitation has, so far, not resulted in any improvement of the situation. It is also shown that scenario-earthquakes developed by disaggregation from the results of a traditional PSHA may not be conservative with respect to energy conservation and should not be used for the design of critical infrastructures without validation. Because the assessment of technical as well as of financial risks associated with potential damages of earthquakes need a risk analysis, current method is based on a probabilistic approach with its unsolved deficiencies.
Traditional deterministic or scenario-based seismic hazard analysis methods provide a reliable and in general robust design basis for applications such as the design of critical infrastructures, especially with systematic sensitivity analyses based on validated phenomenological models. Deterministic seismic hazard analysis incorporates uncertainties in the safety factors. These factors are derived from experience as well as from expert judgment. Deterministic methods associated with high safety factors may lead to too conservative results, especially if applied for generally short-lived civil structures. Scenarios used in deterministic seismic hazard analysis have a clear physical basis. They are related to seismic sources discovered by geological, geomorphologic, geodetic and seismological investigations or derived from historical references. Scenario-based methods can be expanded for risk analysis applications with an extended data analysis providing the frequency of seismic events. Such an extension provides a better informed risk model that is suitable for risk-informed decision making. 相似文献
By definition, a crisis is a situation that requires assistance to be managed. Hence, response to a crisis involves the merging
of local and non-local emergency response personnel. In this situation, it is critical that each participant: (1) know the
roles and responsibilities of each of the other participants; (2) know the capabilities of each of the participants; and (3)
have a common basis for action. For many types of natural disasters, this entails having a common operational picture of the unfolding events, including detailed information on the weather, both current and forecasted, that may impact on either
the emergency itself or on response activities. The Consequences Assessment Tool Set (CATS) is a comprehensive package of
hazard prediction models and casualty and damage assessment tools that provides a linkage between a modeled or observed effect
and the attendant consequences for populations, infrastructure, and resources, and, hence, provides the common operational
picture for emergency response. The Operational Multiscale Environment model with Grid Adaptivity (OMEGA) is an atmospheric
simulation system that links the latest methods in computational fluid dynamics and high-resolution gridding technologies
with numerical weather prediction to provide specific weather analysis and forecast capability that can be merged into the
geographic information system framework of CATS. This paper documents the problem of emergency response as an end-to-end system
and presents the integrated CATS–OMEGA system as a prototype of such a system that has been used successfully in a number
of different situations. 相似文献
Quantitative sinkhole hazard assessments in karst areas allow calculation of the potential sinkhole risk and the performance
of cost-benefit analyses. These estimations are of practical interest for planning, engineering, and insurance purposes. The
sinkhole hazard assessments should include two components: the probability of occurrence of sinkholes (sinkholes/km2 year) and the severity of the sinkholes, which mainly refers to the subsidence mechanisms (progressive passive bending or
catastrophic collapse) and the size of the sinkholes at the time of formation; a critical engineering design parameter. This
requires the compilation of an exhaustive database on recent sinkholes, including information on the: (1) location, (2) chronology
(precise date or age range), (3) size, and (4) subsidence mechanisms and rate. This work presents a hazard assessment from
an alluvial evaporite karst area (0.81 km2) located in the periphery of the city of Zaragoza (Ebro River valley, NE Spain). Five sinkholes and four locations with features
attributable to karstic subsidence where identified in an initial investigation phase providing a preliminary probability
of occurrence of 0.14 sinkholes/km2 year (11.34% in annual probability). A trenching program conducted in a subsequent investigation phase allowed us to rule
out the four probable sinkholes, reducing the probability of occurrence to 0.079 sinkholes/km2 year (6.4% in annual probability). The information on the severity indicates that collapse sinkholes 10–15 m in diameter
may occur in the area. A detailed study of the deposits and deformational structures exposed by trenching in one of the sinkholes
allowed us to infer a modern collapse sinkhole approximately 12 m in diameter and with a vertical throw of 8 m. This collapse
structure is superimposed on a subsidence sinkhole around 80 m across that records at least 1.7 m of synsedimentary subsidence.
Trenching, in combination with dating techniques, is proposed as a useful methodology to elucidate the origin of depressions
with uncertain diagnosis and to gather practical information with predictive utility about particular sinkholes in alluvial
karst settings: precise location, subsidence mechanisms and magnitude, and timing and rate of the subsidence episodes. 相似文献
Agriculture in the southern Great Plains of Canada has been particularly vulnerable to prolonged episodes of drought. Using
climate data and a precipitation minus potential evapotranspiration index, the extent of the region’s exposure to drought
is examined. Between 1914 and 1917, the Dry Belt was particularly vulnerable to drought, whereas after 1928, a much larger
region known as the Palliser Triangle covering most of southern Alberta and Saskatchewan was much more exposed to drought.
These droughts provoked major institutional adaptation, in particular the establishment of the Special Areas Board by the
Government of Alberta, and the creation of the Prairie Farm Rehabilitation Administration by the Government of Canada. Both
organizations have proved to be relatively permanent public adaptations to the natural hazard of drought in the region. Moreover,
these earlier experiences with prolonged drought as well as institution-building may be of value in helping the residents
of the Palliser Triangle adapt to predicted climate changes in the future as well as anticipate some of the barriers to effective
institutional adaptation. 相似文献
Concern for natural hazard-triggered technological disasters (Natech disasters) in densely populated and industrialized areas
is growing. Residents living in urban areas subject to high natural hazard risk are often unaware of the potential for secondary
disasters such as hazardous materials releases from neighboring industrial facilities, chemical storage warehouses or other
establishments housing hazardous materials. Lessons from previous disasters, such as the Natech disaster during the Kocaeli
earthquake in Turkey in 1999 call for the need to manage low frequency/high consequence events, particularly in today’s densely
populated areas. However, there is little guidance available on how local governments and communities can assess Natech risk.
To add to the problem, local governments often do not have the human or economic resources or expertise to carry out detailed
risk assessments. In this article, we propose a methodology for preliminary assessment of Natech risk in urban areas. The
proposed methodology is intended for use by local government officials in consultation with the public. The methodology considers
possible interactions between the various systems in the urban environment: the physical infrastructure (e.g., industrial
plants, lifeline systems, critical facilities), the community (e.g., population exposed), the natural environment (e.g., delicate
ecosystems, river basins), and the risk and emergency management systems (e.g., structural and nonstructural measures). Factors
related to vulnerability and hazard are analyzed and qualitative measures are recommended. Data from hazardous materials releases
during the Kocaeli, Turkey earthquake of August 17, 1999 are used as a case study to demonstrate the applicability of the
methodology. Limitations of the proposed methodology are discussed as well as future research needs.