首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Performance‐based engineering (PBE) methodologies allow for the design of more reliable earthquake‐resistant structures. Nonetheless, to implement PBE techniques, accurate finite element models of critical components are needed. With these objectives in mind, initially, we describe an experimental study on the seismic behaviour of both beam‐to‐column (BTC) and column‐base (CB) joints made of high‐strength steel S590 circular columns filled with concrete. These joints belonged to moment‐resisting frames (MRFs) that constituted the lateral‐force‐resisting system of an office building. BTC joints were conceived as rigid and of partial strength, whereas CB joints were designed as rigid and of full strength. Tests on a BTC joint composed of an S275 steel composite beam and high‐strength steel concrete‐filled tubes were carried out. Moreover, two seismic CB joints were tested with stiffeners welded to the base plate and anchor bolts embedded in the concrete foundation as well as where part of a column was embedded in the foundation with no stiffeners. A test programme was carried out with the aim of characterising these joints under monotonic, cyclic and random loads. Experimental results are presented by means of both force–interstory drift ratio and moment–rotation relationships. The outcomes demonstrated the adequacy of these joints to be used for MRFs of medium ductility class located in zones of moderate seismic hazard. Then, a numerical calibration of the whole joint subassemblies was successfully accomplished. Finally, non‐linear time‐history analyses performed on 2D MRFs provided useful information on the seismic behaviour of relevant MRFs. Copyright © 2015 John Wiley & Sons, Ltd.  相似文献   

2.
In this paper, we discuss the effects of anomalous out‐of‐plane bodies in two‐dimensional (2D) borehole‐to‐surface electrical resistivity tomography with numerical resistivity modelling and synthetic inversion tests. The results of the two groups of synthetic resistivity model tests illustrate that anomalous bodies out of the plane of interest have an effect on two‐dimensional inversion and that the degree of influence of out‐of‐plane body on inverted images varies. The different influences are derived from two cases. One case is different resistivity models with the same electrode array, and the other case is the same resistivity model with different electrode arrays. Qualitative interpretation based on the inversion tests shows that we cannot find a reasonable electrode array to determine the best inverse solution and reveal the subsurface resistivity distribution for all types of geoelectrical models. Because of the three‐dimensional effect arising from neighbouring anomalous bodies, the qualitative interpretation of inverted images from the two‐dimensional inversion of electrical resistivity tomography data without prior information can be misleading. Two‐dimensional inversion with drilling data can decrease the three‐dimensional effect. We employed two‐ and three‐dimensional borehole‐to‐surface electrical resistivity tomography methods with a pole–pole array and a bipole–bipole array for mineral exploration at Abag Banner and Hexigten Banner in Inner Mongolia, China. Different inverse schemes were carried out for different cases. The subsurface resistivity distribution obtained from the two‐dimensional inversion of the field electrical resistivity tomography data with sufficient prior information, such as drilling data and other non‐electrical data, can better describe the actual geological situation. When there is not enough prior information to carry out constrained two‐dimensional inversion, the three‐dimensional electrical resistivity tomography survey is the better choice.  相似文献   

3.
Reliable estimation of the volume and timing of snowmelt runoff is vital for water supply and flood forecasting in snow‐dominated regions. Snowmelt is often simulated using temperature‐index (TI) models due to their applicability in data‐sparse environments. Previous research has shown that a modified‐TI model, which uses a radiation‐derived proxy temperature instead of air temperature as its surrogate for available energy, can produce more accurate snow‐covered area (SCA) maps than a traditional TI model. However, it is unclear whether the improved SCA maps are associated with improved snow water equivalent (SWE) estimation across the watershed or improved snowmelt‐derived streamflow simulation. This paper evaluates whether a modified‐TI model produces better streamflow estimates than a TI model when they are used within a fully distributed hydrologic model. It further evaluates the performance of the two models when they are calibrated using either point SWE measurements or SCA maps. The Senator Beck Basin in Colorado is used as the study site because its surface is largely bedrock, which reduces the role of infiltration and emphasizes the role of the SWE pattern on streamflow generation. Streamflow is simulated using both models for 6 years. The modified‐TI model produces more accurate streamflow estimates (including flow volume and peak flow rate) than the TI model, likely because the modified‐TI model better reproduces the SWE pattern across the watershed. Both models also produce better performance when calibrated with SCA maps instead of point SWE data, likely because the SCA maps better constrain the space‐time pattern of SWE.  相似文献   

4.
In the application of a conventional common‐reflection‐surface (CRS) stack, it is well‐known that only one optimum stacking operator is determined for each zero‐offset sample to be simulated. As a result, the conflicting dip situations are not taken into account and only the most prominent event contributes to any a particular stack sample. In this paper, we name this phenomenon caused by conflicting dip problems as ‘dip discrimination phenomenon’. This phenomenon is not welcome because it not only leads to the loss of weak reflections and tips of diffractions in the final zero‐offset‐CRS stacked section but also to a deteriorated quality in subsequent migration. The common‐reflection‐surface stack with the output imaging scheme (CRS‐OIS) is a novel technique to implement a CRS stack based on a unified Kirchhoff imaging approach. As far as dealing with conflicting dip problems is concerned, the CRS‐OIS is a better option than a conventional CRS stack. However, we think the CRS‐OIS can do more in this aspect. In this paper, we propose a workflow to handle the dip discrimination phenomenon based on a cascaded implementation of prestack time migration, CRS‐OIS and prestack time demigration. Firstly, a common offset prestack time migration is implemented. Then, a CRS‐OIS is applied to the time‐migrated common offset gather. Afterwards, a prestack time demigration is performed to reconstruct each unmigrated common offset gather with its reflections being greatly enhanced and diffractions being well preserved. Compared with existing techniques dealing with conflicting dip problems, the technique presented in this paper preserves most of the diffractions and accounts for reflections from all possible dips properly. More importantly, both the post‐stacked data set and prestacked data set can be of much better quality after the implementation of the presented scheme. It serves as a promising alternative to other techniques except that it cannot provide the typical CRS wavefield attributes. The numerical tests on a synthetic Marmousi data set and a real 2D marine data set demonstrated its effectiveness and robustness.  相似文献   

5.
Major earthquakes (i.e., mainshocks) typically trigger a sequence of lower magnitude events clustered both in time and space. Recent advances of seismic hazard analysis stochastically model aftershock occurrence (given the main event) as a nonhomogeneous Poisson process with rate that decays in time as a negative power law. Risk management in the post‐event emergency phase has to deal with this short‐term seismicity. In fact, because the structural systems of interest might have suffered some damage in the mainshock, possibly worsened by damaging aftershocks, the failure risk may be large until the intensity of the sequence reduces or the structure is repaired. At the state‐of‐the‐art, the quantitative assessment of aftershock risk is aimed at building tagging, that is, to regulate occupancy. The study, on the basis of age‐dependent stochastic processes, derived closed‐form approximations for the aftershock reliability of simple nonevolutionary elastic‐perfectly‐plastic damage‐cumulating systems, conditional on different information about the structure. Results show that, in the case hypotheses apply, the developed models may represent a basis for handy tools enabling risk‐informed tagging by stakeholders and decision makers. Copyright © 2013 John Wiley & Sons, Ltd.  相似文献   

6.
It is important to include the viscous effect in seismic numerical modelling and seismic migration due to the ubiquitous viscosity in an actual subsurface medium. Prestack reverse‐time migration (RTM) is currently one of the most accurate methods for seismic imaging. One of the key steps of RTM is wavefield forward and backward extrapolation and how to solve the wave equation fast and accurately is the essence of this process. In this paper, we apply the time‐space domain dispersion‐relation‐based finite‐difference (FD) method for visco‐acoustic wave numerical modelling. Dispersion analysis and numerical modelling results demonstrate that the time‐space domain FD method has great accuracy and can effectively suppress numerical dispersion. Also, we use the time‐space domain FD method to solve the visco‐acoustic wave equation in wavefield extrapolation of RTM and apply the source‐normalized cross‐correlation imaging condition in migration. Improved imaging has been obtained in both synthetic and real data tests. The migration result of the visco‐acoustic wave RTM is clearer and more accurate than that of acoustic wave RTM. In addition, in the process of wavefield forward and backward extrapolation, we adopt adaptive variable‐length spatial operators to compute spatial derivatives to significantly decrease computing costs without reducing the accuracy of the numerical solution.  相似文献   

7.
Late Triassic A‐type granites are identified in this study in Sarudik, SW Sumatra. We present new data on zircon U–Pb geochronology, whole‐rock major and trace elements and Sr‐Nd‐Hf isotope geochemistry, aiming to study their petrogenesis and tectonic implications. LA‐ICP‐MS U–Pb dating of zircon separated from one biotite monzogranite sample yields a concordia age of 222.6 ±1.0 Ma, indicating a Late Triassic magmatic event. The studied granites are classified as weakly peralumious, high‐K calc‐alkaline granites. They exhibit high SiO2, K2O + Na2O, FeO/(FeO + MgO) and Ga/Al ratios and low Al2O3, CaO, MgO, P2O5 and TiO2 contents, with enrichment of Rb, Th and U and depletion of Ba, Sr, P and Eu, showing the features of A‐type granites. The granites have zircon εHf(t) values from ?4.6 to ?0.4 and whole‐rock εNd(t) values from ?5.51 to ?4.98, with Mesoproterozoic TDM2 ages (1278–1544 Ma) for both Hf and Nd isotopes. Geochemical and isotopic data suggest that the source of these A‐type granites is the Mesoproterozoic continental crust, without significant incorporation of mantle‐derived component, and their formation is controlled by subsequent fractional crystallization. The Sarudik A‐type granites are further assigned to A2‐type formed in post‐collisional environment. Combined with previous knowledge on the western SE Asia tectonic evolution, we conclude that the formation of the Late Triassic A‐type granites is related to the post‐collisional extension induced by the crustal thickening, gravitational collapse, and asthenosphere upwelling following the collision between the Sibumasu and the East Malaya Block.  相似文献   

8.
The assessment of earthquake loss often requires the definition of a relation between a measure of damage and a quantity of loss, usually achieved through the employment of a damage‐to‐loss model. These models are frequently characterized by a large variability, which inevitably increases the uncertainty in the vulnerability assessment and earthquake loss estimation. This study provides an insight on the development of damage‐to‐loss functions for moment‐frame reinforced concrete buildings through an analytical methodology. Tri‐dimensional finite element models of existing reinforced concrete buildings were subjected to a number of ground motion records compatible with the seismicity in the region of interest, through nonlinear dynamic analysis. These results were used to assess, for a number of damage states, the probability distribution of loss ratio, taking into consideration member damage and different repair techniques, as well as to derive sets of fragility functions. Then, a vulnerability model (in terms of the ratio of cost of repair to cost of replacement, conditional on the level of ground shaking intensity) was derived and compared with the vulnerability functions obtained through the combination of various damage‐to‐loss models with the set of fragility functions developed herein. In order to provide realistic estimates of economic losses due to seismic action, a comprehensive study on repair costs using current Portuguese market values was also carried out. The results of this study highlight important issues in the derivation of vulnerability functions, which are a fundamental component for an adequate seismic risk assessment. © 2015 The Authors. Earthquake Engineering & Structural Dynamics published by John Wiley & Sons Ltd.  相似文献   

9.
This study presents single‐objective and multi‐objective particle swarm optimization (PSO) algorithms for automatic calibration of Hydrologic Engineering Center‐ Hydrologic Modeling Systems rainfall‐runoff model of Tamar Sub‐basin of Gorganroud River Basin in north of Iran. Three flood events were used for calibration and one for verification. Four performance criteria (objective functions) were considered in multi‐objective calibration where different combinations of objective functions were examined. For comparison purposes, a fuzzy set‐based approach was used to determine the best compromise solutions from the Pareto fronts obtained by multi‐objective PSO. The candidate parameter sets determined from different single‐objective and multi‐objective calibration scenarios were tested against the fourth event in the verification stage, where the initial abstraction parameters were recalibrated. A step‐by‐step screening procedure was used in this stage while evaluating and comparing the candidate parameter sets, which resulted in a few promising sets that performed well with respect to at least three of four performance criteria. The promising sets were all from the multi‐objective calibration scenarios which revealed the outperformance of the multi‐objective calibration on the single‐objective one. However, the results indicated that an increase of the number of objective functions did not necessarily lead to a better performance as the results of bi‐objective function calibration with a proper combination of objective functions performed as satisfactorily as those of triple‐objective function calibration. This is important because handling multi‐objective optimization with an increased number of objective functions is challenging especially from a computational point of view. Copyright © 2012 John Wiley & Sons, Ltd.  相似文献   

10.
The present work focuses on the performance of Fenton, sono‐Fenton, and sono‐photo‐Fenton processes for the oxidation of phenol present in aqueous solution. The effects of H2O2 concentration, Fe2+ concentration, pH, and initial phenol concentration on the oxidation of phenol were studied. The optimum Fe2+ and H2O2 concentrations for the Fenton process were 45 and 800 mg/L, respectively. For the sono‐Fenton process, the optimum Fe2+ and H2O2 concentrations were 30 and 800 mg/L, respectively. The optimal conditions for the sono‐photo‐Fenton process were found to be 20 mg/L of Fe2+ and 700 mg/L of H2O2. The optimum pH was found to be 3 for the processes investigated in the present study. The analysis of results showed that the sono‐photo‐Fenton method reduced the Fe2+ concentration by 30–50% and the H2O2 concentration by 12.5%. It was found that the sono‐photo‐Fenton technique showed better performance than the Fenton and sono‐Fenton processes for the oxidation of phenol. A lumped kinetic model was used to predict the chemical oxygen demand reduction and the model was found to fit the data.  相似文献   

11.
A network of 30 standalone snow monitoring stations was used to investigate the snow cover distribution, snowmelt dynamics, and runoff generation during two rain‐on‐snow (ROS) events in a 40 km2 montane catchment in the Black Forest region of southwestern Germany. A multiple linear regression analysis using elevation, aspect, and land cover as predictors for the snow water equivalent (SWE) distribution within the catchment was applied on an hourly basis for two significant ROS flood events that occurred in December 2012. The available snowmelt water, liquid precipitation, as well as the total retention storage of the snow cover were considered in order to estimate the amount of water potentially available for the runoff generation. The study provides a spatially and temporally distributed picture of how the two observed ROS floods developed in the catchment. It became evident that the retention capacity of the snow cover is a crucial mechanism during ROS. It took several hours before water was released from the snowpack during the first ROS event, while retention storage was exceeded within 1 h from the start of the second event. Elevation was the most important terrain feature. South‐facing terrain contributed more water for runoff than north‐facing slopes, and only slightly more runoff was generated at open compared to forested areas. The results highlight the importance of snowmelt together with liquid precipitation for the generation of flood runoff during ROS and the large temporal and spatial variability of the relevant processes. Copyright © 2015 John Wiley & Sons, Ltd.  相似文献   

12.
The role of masonry infills in the seismic behavior of reinforced concrete buildings has been widely studied in terms of their strength and stiffness contribution in the in‐plane (IP) direction, while fewer studies have been carried out on their response and modeling in the out‐of‐plane (OOP) direction. In this paper, the state of the art in code and literature provisions regarding infills' OOP capacity and seismic demand is presented, together with a review of the experimental tests that have been carried out to investigate infills' OOP behavior and the effects of IP‐OOP interaction. This review aims to collect an experimental database that is used to evaluate the effectiveness of literature and code provisions and to propose a semiempirical approach both for predicting infills' OOP strength, stiffness, and displacement capacity and for modeling the effects of IP displacement demand on OOP behavior and vice versa. Then, the state of the art on modeling of infills' OOP behavior and IP‐OOP interaction is presented together with a new macro model based on the proposed formulations and conceived to represent the IP and OOP behavior by taking into account the mutual interaction effects. Finally, the proposed model is used for an example application on two case‐study buildings, showing the effects of taking into account or neglecting the IP‐OOP interaction phenomena.  相似文献   

13.
The response of low‐ductility reinforced concrete (RC) frames, designed typically for a non‐seismic region, subjected to two frequencies of base excitations is studied. Five half‐scaled, two‐bay, two‐storey, RC frames, each approximately 5 m wide by 3.3 m high, were subjected to both horizontal and/or vertical base excitations with a frequency of 40 Hz as well as a lower frequency of about 4 Hz (close to the fundamental frequency) using a shake table. The imposed acceleration amplitude ranged from 0.2 to 1.2g. The test results showed that the response characteristics of the structures differed under high‐ and low‐frequency excitations. The frames were able to sustain high‐frequency excitations without damage but were inadequate for low‐frequency excitations, even though the frames exhibited some ductility. Linear‐elastic time‐history analysis can predict reasonably well the structural response under high‐frequency excitations. As the frames were not designed for seismic loads, the reinforcement detailing may not have been adequate, based on the crack pattern observed. The effect of vertical excitation can cause significant additional forces in the columns and moment reversals in the beams. The ‘strong‐column, weak‐beam’ approach for lateral load RC frame design is supported by experimental observations. Copyright © 2001 John Wiley & Sons, Ltd.  相似文献   

14.
A series of large‐scale real‐time hybrid simulations (RTHSs) are conducted on a 0.6‐scale 3‐story steel frame building with magneto‐rheological (MR) dampers. The lateral force resisting system of the prototype building for the study consists of moment resisting frames and damped brace frames (DBFs). The experimental substructure for the RTHS is the DBF with the MR dampers, whereas the remaining structural components of the building including the moment resisting frame and gravity frames are modeled via a nonlinear analytical substructure. Performing RTHS with an experimental substructure that consists of the complete DBF enables the effects of member and connection component deformations on system and damper performance to be accurately accounted for. Data from these tests enable numerical simulation models to be calibrated, provide an understanding and validation of the in‐situ performance of MR dampers, and a means of experimentally validating performance‐based seismic design procedures for real structures. The details of the RTHS procedure are given, including the test setup, the integration algorithm, and actuator control. The results from a series of RTHS are presented that includes actuator control, damper behavior, and the structural response for different MR control laws. The use of the MR dampers is experimentally demonstrated to reduce the response of the structure to strong ground motions. Comparisons of the RTHS results are made with numerical simulations. Based on the results of the study, it is concluded that RTHS can be conducted on realistic structural systems with dampers to enable advancements in resilient earthquake resistant design to be achieved. Copyright © 2014 John Wiley & Sons, Ltd.  相似文献   

15.
This paper describes least‐squares reverse‐time migration. The method provides the exact adjoint operator pair for solving the linear inverse problem, thereby enhancing the convergence of gradient‐based iterative linear inversion methods. In this formulation, modified source wavelets are used to correct the source signature imprint in the predicted data. Moreover, a roughness constraint is applied to stabilise the inversion and reduce high‐wavenumber artefacts. It is also shown that least‐squares migration implicitly applies a deconvolution imaging condition. Three numerical experiments illustrate that this method is able to produce seismic reflectivity images with higher resolution, more accurate amplitudes, and fewer artefacts than conventional reverse‐time migration. The methodology is currently feasible in 2‐D and can naturally be extended to 3‐D when computational resources become more powerful.  相似文献   

16.
Reverse‐time migration gives high‐quality, complete images by using full‐wave extrapolations. It is thus not subject to important limitations of other migrations that are based on high‐frequency or one‐way approximations. The cross‐correlation imaging condition in two‐dimensional pre‐stack reverse‐time migration of common‐source data explicitly sums the product of the (forward‐propagating) source and (backward‐propagating) receiver wavefields over all image times. The primary contribution at any image point travels a minimum‐time path that has only one (specular) reflection, and it usually corresponds to a local maximum amplitude. All other contributions at the same image point are various types of multipaths, including prismatic multi‐arrivals, free‐surface and internal multiples, converted waves, and all crosstalk noise, which are imaged at later times, and potentially create migration artefacts. A solution that facilitates inclusion of correctly imaged, non‐primary arrivals and removal of the related artefacts, is to save the depth versus incident angle slice at each image time (rather than automatically summing them). This results in a three‐parameter (incident angle, depth, and image time) common‐image volume that integrates, into a single unified representation, attributes that were previously computed by separate processes. The volume can be post‐processed by selecting any desired combination of primary and/or multipath data before stacking over image time. Separate images (with or without artifacts) and various projections can then be produced without having to remigrate the data, providing an efficient tool for optimization of migration images. A numerical example for a simple model shows how primary and prismatic multipath contributions merge into a single incident angle versus image time trajectory. A second example, using synthetic data from the Sigsbee2 model, shows that the contributions to subsalt images of primary and multipath (in this case, turning wave) reflections are different. The primary reflections contain most of the information in regions away from the salt, but both primary and multipath data contribute in the subsalt region.  相似文献   

17.
In many land seismic situations, the complex seismic wave propagation effects in the near‐surface area, due to its unconsolidated character, deteriorate the image quality. Although several methods have been proposed to address this problem, the negative impact of 3D complex near‐surface structures is still unsolved to a large extent. This paper presents a complete 3D data‐driven solution for the near‐surface problem based on 3D one‐way traveltime operators, which extends our previous attempts that were limited to a 2D situation. Our solution is composed of four steps: 1) seismic wave propagation from the surface to a suitable datum reflector is described by parametrized one‐way propagation operators, with all the parameters estimated by a new genetic algorithm, the self‐adjustable input genetic algorithm, in an automatic and purely data‐driven way; 2) surface‐consistent residual static corrections are estimated to accommodate the fast variations in the near‐surface area; 3) a replacement velocity model based on the traveltime operators in the good data area (without the near‐surface problem) is estimated; 4) data interpolation and surface layer replacement based on the estimated traveltime operators and the replacement velocity model are carried out in an interweaved manner in order to both remove the near‐surface imprints in the original data and keep the valuable geological information above the datum. Our method is demonstrated on a subset of a 3D field data set from the Middle East yielding encouraging results.  相似文献   

18.
Ti‐SBA‐15 materials with Ti incorporated into the framework of SBA‐15 and controllable Ti contents were successfully prepared via a post‐treatment route in supercritical CO2‐ethanol solution, followed by calcination. The resultant Ti‐SBA‐15 materials were characterized by means of different techniques including X‐ray diffraction, X‐ray photoelectron spectroscopy, transmission electron microscopy, IR analysis, and N2 sorption analysis. It was demonstrated that the resultant materials retained a structure similar to that of the parent SBA‐15, and Ti was incorporated into the framework of SBA‐15. At high Ti content, i.e, Si/Ti = 5:1, a TiO2 phase formed and was coated onto the inner surface of SBA‐15 in addition to the incorporation of the Ti in the framework. The BET surface areas of the Ti‐SBA‐15 samples decreased with increasing Ti content. The presence of small amounts of H2O in the reaction medium may have resulted in some TiO2 nanoparticles being uniformly distributed in the pores of the SBA‐15 accompanying the incorporation of Ti in the SBA‐15 framework.  相似文献   

19.
The seismic assessment of the local failure modes in existing masonry buildings is currently based on the identification of the so‐called local mechanisms, often associated with the out‐of‐plane wall behavior, whose stability is evaluated by static force‐based approaches and, more recently, by some displacement‐based proposals. Local mechanisms consist of kinematic chains of masonry portions, often regarded as rigid bodies, with geometric nonlinearity and concentrated nonlinearity in predefined contact regions (unilateral no‐tension behavior, possible sliding with friction). In this work, the dynamic behavior of local mechanisms is simulated through multi‐body dynamics, to obtain the nonlinear response with efficient time history analyses that directly take into account the characteristics of the ground motion. The amplification/filtering effects of the structure are considered within the input motion. The proposed approach is validated with experimental results of two full‐scale shaking‐table tests on stone masonry buildings: a sacco‐stone masonry façade tested at Laboratório Nacional de Engenharia Civil and a two‐storey double‐leaf masonry building tested at European Centre for Training and Research in Earthquake Engineering (EUCENTRE). Copyright © 2015 John Wiley & Sons, Ltd.  相似文献   

20.
An indigenous bacterial strain of Delftia sp. capable of degrading 2,4‐dicholorophenol and an indigenous bacterial community that degrades 2,4,6‐trichlorophenol (TCP) were employed to inoculate continuous down‐flow fixed‐bed reactors. Continuous‐reactors were constructed from PVC employing hollow PVC cylinders as support material. Synthetic wastewater was prepared by dissolving the corresponding chlorophenol in non‐sterile groundwater. Biodegradation was evaluated by spectrophotometry, chloride release, GC, and microbial growth. Detoxification was evaluated by using Daphnia magna as test organism. Delftia sp. was able to remove an average of 95.6% of DCP. Efficiency in terms of chemical oxygen demand (COD) was of 88.9%. The indigenous bacterial community that degrades TCP reached an average efficiency of 96.5 and 91.6% in terms of compound and COD removal, respectively. In both cases stoichiometric removal of chloride and detoxification was achieved. When synthetic wastewater feed was cut off for 7 days, both reactors showed a fast recovery after inflow restarting, reaching average outlet concentration values within 36 h. The promising behavior of the microorganisms and the low cost of the reactors tested allow us to suggest their possible application to remediation processes.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号