首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
在研究分析地址模型的基础上,建立了存储标准地址数据集的标准地址库和自定义的地址匹配规则库,提出了一种基于规则的模糊中文地址编码方法。该方法在依据标准地址库分词的同时,也沿着自定义的地址匹配规则进行推理,从而缩小了下次分词所用到的目标数据集,提高了系统执行效率。另外,通过借助构建的规则树与歧义栈,提高了文中定义的两类模糊地址匹配的成功率。最后,基于该算法建立了一个地理编码原型系统,并利用经济普查项目中的相关数据对算法的可用性进行了验证。  相似文献   

2.
The techniques of fuzzy logic and Monte Carlo simulation are combined to address two incompatible types of uncertainty present in most natural resource data: thematic classification uncertainty and variance in unclassified continuously distributed data. The resultant model of uncertainty is applied to an infinite slope stability model using data from Louise Island, British Columbia. Results are summarized so as to answer forestry decision support queries. The proposed model of uncertainty in resource data analysis is found to have utility in combining different types of uncertainty, and efficiently utilizing available metadata. Integration of uncertainty data models with visualization tools is considered a necessary prerequisite to effective implementation in decision support systems.  相似文献   

3.
Spatial data uncertainty models (SDUM) are necessary tools that quantify the reliability of results from geographical information system (GIS) applications. One technique used by SDUM is Monte Carlo simulation, a technique that quantifies spatial data and application uncertainty by determining the possible range of application results. A complete Monte Carlo SDUM for generalized continuous surfaces typically has three components: an error magnitude model, a spatial statistical model defining error shapes, and a heuristic that creates multiple realizations of error fields added to the generalized elevation map. This paper introduces a spatial statistical model that represents multiple statistics simultaneously and weighted against each other. This paper's case study builds a SDUM for a digital elevation model (DEM). The case study accounts for relevant shape patterns in elevation errors by reintroducing specific topological shapes, such as ridges and valleys, in appropriate localized positions. The spatial statistical model also minimizes topological artefacts, such as cells without outward drainage and inappropriate gradient distributions, which are frequent problems with random field-based SDUM. Multiple weighted spatial statistics enable two conflicting SDUM philosophies to co-exist. The two philosophies are ‘errors are only measured from higher quality data’ and ‘SDUM need to model reality’. This article uses an automatic parameter fitting random field model to initialize Monte Carlo input realizations followed by an inter-map cell-swapping heuristic to adjust the realizations to fit multiple spatial statistics. The inter-map cell-swapping heuristic allows spatial data uncertainty modelers to choose the appropriate probability model and weighted multiple spatial statistics which best represent errors caused by map generalization. This article also presents a lag-based measure to better represent gradient within a SDUM. This article covers the inter-map cell-swapping heuristic as well as both probability and spatial statistical models in detail.  相似文献   

4.
We analysed the sensitivity of a decision tree derived forest type mapping to simulated data errors in input digital elevation model (DEM), geology and remotely sensed (Landsat Thematic Mapper) variables. We used a stochastic Monte Carlo simulation model coupled with a one‐at‐a‐time approach. The DEM error was assumed to be spatially autocorrelated with its magnitude being a percentage of the elevation value. The error of categorical geology data was assumed to be positional and limited to boundary areas. The Landsat data error was assumed to be spatially random following a Gaussian distribution. Each layer was perturbed using its error model with increasing levels of error, and the effect on the forest type mapping was assessed. The results of the three sensitivity analyses were markedly different, with the classification being most sensitive to the DEM error, than to the Landsat data errors, but with only a limited sensitivity to the geology data error used. A linear increase in error resulted in non‐linear increases in effect for the DEM and Landsat errors, while it was linear for geology. As an example, a DEM error of as small as ±2% reduced the overall test accuracy by more than 2%. More importantly, the same uncertainty level has caused nearly 10% of the study area to change its initial class assignment at each perturbation, on average. A spatial assessment of the sensitivities indicates that most of the pixel changes occurred within those forest classes expected to be more sensitive to data error. In addition to characterising the effect of errors on forest type mapping using decision trees, this study has demonstrated the generality of employing Monte Carlo analysis for the sensitivity and uncertainty analysis of categorical outputs that have distinctive characteristics from that of numerical outputs.  相似文献   

5.
ABSTRACT

The focus of this work is general methods for prioritization or screening of project sites based on the favorability of multiple spatial criteria. We present a threshold-based transformation of each underlying spatial favorability factor into a continuous scale with a common favorability interpretation across all criteria. We compare several methods of computing site favorability and propagating uncertainty from the data to the favorability metrics. Including uncertainty allows decision makers to determine if seeming differences among sites are significant. We address uncertainty using Taylor series approximations and analytical distributions, which are compared to computationally intensive Monte Carlo simulations. Our methods are applied to siting direct-use geothermal energy projects in the Appalachian Basin, where our knowledge about any particular site is limited, yet sufficient data exist to estimate favorability. We consider four factors that contribute to site favorability: the thermal resource described by the depth to 80°C rock, natural reservoir productivity described by rock permeability and thickness, potential for induced seismicity, and the estimated cost of surface infrastructure for heat distribution. Those factors are combined in three ways. We develop favorability uncertainty propagation and sensitivity analysis methods. All methods are general and can be applied to other multi-criteria spatial screening problems.  相似文献   

6.
利用在线地理编码API解决海量中文地址快速编码问题,在此基础上,利用简单的规则对编码结果进行清洗、标记,最后通过基于系统聚类与随机森林的分类优化模型,将多平台编码结果分类处理、优化。利用广州市盗窃案件地址对模型进行训练与验证,结果表明:相比未处理的编码结果,经模型优化过的编码结果整体位置误差距离减小。高德的地理编码服务有着最好的编码质量,但训练样本的高德编码误差均值仍高达590.43 m,经模型优化后,样本的编码误差均值降至173.73 m,验证样本编码误差均值由554.88 m(高德)降至180.04 m,降低了67.49%,高德90.08%的异常编码结果被清洗优化。对于训练样本与验证样本,模型优化效果相似;对于地址类型不同的案件、位于市区与市郊的案件,模型优化效果相似,说明模型具有一定普适性。该模型能够方便快捷地将海量社会经济信息转化为空间数据,提高编码精度,为地理大数据的研究提供更好的数据支持。  相似文献   

7.
This article applies error propagation in a Monte Carlo simulation for a spatial-based fuzzy logic multi-criteria evaluation (MCE) in order to investigate the output uncertainty created by the input data sets and model structure. Six scenarios for quantifying uncertainty are reviewed. Three scenarios are progressively more complex in defining observational data (attribute uncertainty); while three other scenarios include uncertainty in observational data (position of boundaries between map units), weighting of evidence (fuzzy membership assignment), and evaluating changes in the MCE model (fuzzy logic operators). A case study of petroleum exploration in northern South America is used. Despite the resources and time required, the best estimate of input uncertainty is that based on expert-defined values. Uncertainties for fuzzy membership assignment and boundary transition zones do not affect the results as much as the attribute assignment uncertainty. The MCE fuzzy logic operator uncertainty affects the results the most. Confidence levels of 95% and 60% are evaluated with threshold values of 0.7 and 0.5 and show that accepting more uncertainty in the results increases the total area available for decision-making. Threshold values and confidence levels should be predetermined, although a series of combinations may yield the best decision-making support.  相似文献   

8.
One of the uses of geostatistical conditional simulation is as a tool in assessing the spatial uncertainty of inputs to the Monte Carlo method of system uncertainty analysis. Because the number of experimental data in practical applications is limited, the geostatistical parameters used in the simulation are themselves uncertain. The inference of these parameters by maximum likelihood allows for an easy assessment of this estimation uncertainty which, in turn, may be included in the conditional simulation procedure. A case study based on transmissivity data is presented to show the methodology whereby both model selection and parameter inference are solved by maximum likelihood.  相似文献   

9.
Data recorded by the Italian Telemetered Seismic Network (ITSN) of the Istituto Nazionale di Geofisica (ING) have been widely used in recent years to image slab structures and to find evidence for active processes along the Italian Peninsula. However, the use of seismic data for geostructural purposes may be affected by the well-known trade-off between earthquake location and seismic-velocity parameters. Furthermore, the confidence ellipse predicted by standard procedures may be inadequate for the representation of the probable error of a computed localization. This paper evaluates the probable errors on the hypocentre determinations of the seismic events recorded by the ITSN, using a Monte Carlo method.
We compute synthetic arrival times using a 1-D velocity model appropriate as an average for the Italian area. The hypocentres used are all those recorded by the ITSN during the period January 1992 to March 1994 (1972 events). Station locations are those of the current ITSN configuration. The synthetic arrival times are perturbed with a Gaussian distribution of errors and input to ING's standard hypocentral location procedure, but using crustal velocities differing by 10 per cent from those used to generate them. Each simulation is repeated at least 30 times. Average absolute shifts of hypocentres are assessed in grid cells of linear dimension 33 km covering the whole Italian region.
For regions within the ITSN, shifts are typically 5–10 km in location and up to 20 km in depth. However, for offshore and coastal regions, they are much greater: 50 km or more in both location and depth (far exceeding the equivalent uncertainties quoted by ING bulletins). Possible consequences of this are highlighted by producing a cross-section of subcrustal hypocentres from the Adriatic to the Tyrrhenian Sea, where the large uncertainty in depth precludes any confident interpretation of dipping tectonic features.  相似文献   

10.
The paper presents a computationally efficient meta-modeling approach to spatially explicit uncertainty and sensitivity analysis in a cellular automata (CA) urban growth and land-use simulation model. The uncertainty and sensitivity of the model parameters are approximated using a meta-modeling method called polynomial chaos expansion (PCE). The parameter uncertainty and sensitivity measures obtained with PCE are compared with traditional Monte Carlo simulation results. The meta-modeling approach was found to reduce the number of model simulations necessary to arrive at stable sensitivity estimates. The quality of the results is comparable to the full-order modeling approach, which is computationally costly. The study shows that the meta-modeling approach can significantly reduce the computational effort of carrying out spatially explicit uncertainty and sensitivity analysis in the application of spatio-temporal models.  相似文献   

11.
This paper explores three theoretical approaches for estimating the degree of correctness to which the accuracy figures of a gridded Digital Elevation Model (DEM) have been estimated depending on the number of checkpoints involved in the assessment process. The widely used average‐error statistic Mean Square Error (MSE) was selected for measuring the DEM accuracy. The work was focused on DEM uncertainty assessment using approximate confidence intervals. Those confidence intervals were constructed both from classical methods which assume a normal distribution of the error and from a new method based on a non‐parametric approach. The first two approaches studied, called Chi‐squared and Asymptotic Student t, consider a normal distribution of the residuals. That is especially true in the first case. The second case, due to the asymptotic properties of the t distribution, can perform reasonably well with even slightly non‐normal residuals if the sample size is large enough. The third approach developed in this article is a new method based on the theory of estimating functions which could be considered much more general than the previous two cases. It is based on a non‐parametric approach where no particular distribution is assumed. Thus, we can avoid the strong assumption of distribution normality accepted in previous work and in the majority of current standards of positional accuracy. The three approaches were tested using Monte Carlo simulation for several populations of residuals generated from originally sampled data. Those original grid DEMs, considered as ground data, were collected by means of digital photogrammetric methods from seven areas displaying differing morphology employing a 2 by 2 m sampling interval. The original grid DEMs were subsampled to generate new lower‐resolution DEMs. Each of these new DEMs was then interpolated to retrieve its original resolution using two different procedures. Height differences between original and interpolated grid DEMs were calculated to obtain residual populations. One interpolation procedure resulted in slightly non‐normal residual populations, whereas the other produced very non‐normal residuals with frequent outliers. Monte Carlo simulations allow us to report that the estimating function approach was the most robust and general of those tested. In fact, the other two approaches, especially the Chi‐squared method, were clearly affected by the degree of normality of the residual population distribution, producing less reliable results than the estimating functions approach. This last method shows good results when applied to the different datasets, even in the case of more leptokurtic populations. In the worst cases, no more than 64–128 checkpoints were required to construct an estimate of the global error of the DEM with 95% confidence. The approach therefore is an important step towards saving time and money in the evaluation of DEM accuracy using a single average‐error statistic. Nevertheless, we must take into account that MSE is essentially a single global measure of deviations, and thus incapable of characterizing the spatial variations of errors over the interpolated surface.  相似文献   

12.
区位理论研究长期忽视作为企业基本属性的注册地址并将其与企业实体位置混淆的做法给区位理论带来挑战和发展机遇。本文回顾地理学、经济学、管理学和法学的有限研究,阐述注册地址作为新区位主体、企业实体与注册地址分离现象以及该现象背景下区位分析的新逻辑等内容。研究表明:注册地址是企业法律关系缔结和实现的基点,具有虚拟性和唯一性的特征,其区位主要受税收政策等制度因素的影响,在不同地区注册的企业面临不同的发展前景,并影响地区的财政收入、形象等。企业实体与注册地址的分离现象已较为普遍,注册地址分布在自贸区、开发区等具有政策优势的各类经济区以及欠发达地区的比例高于企业实体。该现象尽管使企业可以兼顾区位优势和政策优势,促进市场活力,但对地区发展、政府部门监管以及企业自身也造成了广泛的负面影响。企业实体与注册地址存在分离现象促使传统“企业”的一元区位体系调整为“企业实体-注册地址”的二元区位体系,而企业实体与注册地址之间的“黏力”使得未来区位理论研究不仅要辨析区位主体是企业实体还是注册地址,而且在分析区位影响、区位因子、区位迁移等内容时要区分两者的位置关系(重合或分离),意识到不同关系下彼此间的相互影响。  相似文献   

13.
Analyses of criminals' travel patterns can provide significant suggestions to improve crime management. This study extends the investigation of criminals' travel behavior from journey‐to‐crime to journey‐after‐crime. Moreover, new methods are developed to examine the spatial patterns of location pairs when restricted by the underlying geographical process. The methods are employed to investigate criminals' journey‐after‐auto‐theft in the city of Buffalo, New York. The analyses reveal that auto thieves' trips from vehicle‐theft locations to the corresponding vehicle‐recovery locations are local in nature. The travel distances are significantly shorter than the randomly simulated trips; the travel directions are biased from the random directions as well.  相似文献   

14.
Rothermel's model is the most widely used fire behaviour model in wildland fire research and management. It is a complex model that considers 17 input variables describing fuel type, fuel moisture, terrain and wind. Uncertainties in the input variables can have a substantial impact on the resulting errors and have to be considered, especially when the results are used in spatial decision making. In this paper it is shown that the analysis of uncertainty propagation can be carried out with the Taylor series method. This method is computationally cheaper than Monte Carlo and offers easy-to-use, preliminary sensitivity estimations.  相似文献   

15.
One of the main objectives of land-use change models is to explore future land-use patterns. Therefore, the issue of addressing uncertainty in land-use forecasting has received an increasing attention in recent years. Many current models consider uncertainty by including a randomness component in their structure. In this paper, we present a novel approach for tuning uncertainty over time, which we refer to as the Time Monte Carlo (TMC) method. The TMC uses a specific range of randomness to allocate new land uses. This range is associated with the transition probabilities from one land use to another. The range of randomness is increased over time so that the degree of uncertainty increases over time. We compare the TMC to the randomness components used in previous models, through a coupled logistic regression-cellular automata model applied for Wallonia (Belgium) as a case study. Our analysis reveals that the TMC produces results comparable with existing methods over the short-term validation period (2000–2010). Furthermore, the TMC can tune uncertainty on longer-term time horizons, which is an essential feature of our method to account for greater uncertainty in the distant future.  相似文献   

16.
As sea level is projected to rise throughout the twenty-first century due to climate change, there is a need to ensure that sea level rise (SLR) models accurately and defensibly represent future flood inundation levels to allow for effective coastal zone management. Digital elevation models (DEMs) are integral to SLR modelling, but are subject to error, including in their vertical resolution. Error in DEMs leads to uncertainty in the output of SLR inundation models, which if not considered, may result in poor coastal management decisions. However, DEM error is not usually described in detail by DEM suppliers; commonly only the RMSE is reported. This research explores the impact of stated vertical error in delineating zones of inundation in two locations along the Devon, United Kingdom, coastline (Exe and Otter Estuaries). We explore the consequences of needing to make assumptions about the distribution of error in the absence of detailed error data using a 1 m, publically available composite DEM with a maximum RMSE of 0.15 m, typical of recent LiDAR-derived DEMs. We compare uncertainty using two methods (i) the NOAA inundation uncertainty mapping method which assumes a normal distribution of error and (ii) a hydrologically correct bathtub method where the DEM is uniformly perturbed between the upper and lower bounds of a 95% linear error in 500 Monte Carlo Simulations (HBM+MCS). The NOAA method produced a broader zone of uncertainty (an increase of 134.9% on the HBM+MCS method), which is particularly evident in the flatter topography of the upper estuaries. The HBM+MCS method generates a narrower band of uncertainty for these flatter areas, but very similar extents where shorelines are steeper. The differences in inundation extents produced by the methods relate to a number of underpinning assumptions, and particularly, how the stated RMSE is interpreted and used to represent error in a practical sense. Unlike the NOAA method, the HBM+MCS model is computationally intensive, depending on the areas under consideration and the number of iterations. We therefore used the HBM+ MCS method to derive a regression relationship between elevation and inundation probability for the Exe Estuary. We then apply this to the adjacent Otter Estuary and show that it can defensibly reproduce zones of inundation uncertainty, avoiding the computationally intensive step of the HBM+MCS. The equation-derived zone of uncertainty was 112.1% larger than the HBM+MCS method, compared to the NOAA method which produced an uncertain area 423.9% larger. Each approach has advantages and disadvantages and requires value judgements to be made. Their use underscores the need for transparency in assumptions and communications of outputs. We urge DEM publishers to move beyond provision of a generalised RMSE and provide more detailed estimates of spatial error and complete metadata, including locations of ground control points and associated land cover.  相似文献   

17.
针对传统空间数据关联规则挖掘缺乏不确定性处理及度量的局限性,将空间数据的不确定性和空间数据挖掘的不确定性有机结合,初步建立了空间数据关联规则挖掘的不确定性处理模型及度量指标,包括空间数据不确定性的Monte Carlo模拟、基于不确定性空间数据的空间自相关度量和关联规则不确定性度量等,并以我国某地区环境调查数据为例进行验证。  相似文献   

18.
Understanding housing submarket structure is of crucial importance to both public and private agencies. It can also help current and future homeowners make informed decisions on their residential choices. Current research on submarket focuses on comparative analyses of different classification techniques. Few studies, however, have examined the function of spatial contiguity on housing submarket classification. To address this issue, this paper developed a spatially constrained data-driven submarket classification methodology to obtain spatially integrated housing market segments. Specifically, a data-driven model based on principal component analysis and cluster analysis was developed for delineating housing submarkets. Within the model, a number of location attributes were used for principal component analysis, and the geographic locations of houses were also incorporated in the cluster analysis. The performance of this method was compared with other unconstrained data-driven techniques and a priori classifications using three measurements: substitutability, spatial integrity, and similarity. Results indicate that spatially contiguous submarkets can be obtained without compromising housing hedonic model accuracy and attribute homogeneity.  相似文献   

19.
20.
The proliferation of digital cameras and the growing practice of online photo sharing using social media sites such as Flickr have resulted in huge volumes of geotagged photos available on the Web. Based on users' traveling preferences elicited from their travel experiences exposed on social media sites by sharing geotagged photos, we propose a new method for recommending tourist locations that are relevant to users (i.e., personalization) in the given context (i.e., context awareness). We obtain user-specific travel preferences from his/her travel history in one city and use these to recommend tourist locations in another city. Our technique is illustrated on a sample of publicly available Flickr dataset containing photos taken in various cities of China. Results show that our context-aware personalized method is able to predict tourists' preferences in a new or unknown city more precisely and generate better recommendations compared to other state-of-the-art landmark recommendation methods.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号