首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
In the context of OpenStreetMap (OSM), spatial data quality, in particular completeness, is an essential aspect of its fitness for use in specific applications, such as planning tasks. To mitigate the effect of completeness errors in OSM, this study proposes a methodological framework for predicting by means of OSM urban areas in Europe that are currently not mapped or only partially mapped. For this purpose, a machine learning approach consisting of artificial neural networks and genetic algorithms is applied. Under the premise of existing OSM data, the model estimates missing urban areas with an overall squared correlation coefficient (R 2) of 0.589. Interregional comparisons of European regions confirm spatial heterogeneity in the model performance, whereas the R 2 ranges from 0.129 up to 0.789. These results show that the delineation of urban areas by means of the presented methodology depends strongly on location.  相似文献   

2.
The vast accumulation of environmental data and the rapid development of geospatial visualization and analytical techniques make it possible for scientists to solicit information from local citizens to map spatial variation of geographic phenomena. However, data provided by citizens (referred to as citizen data in this article) suffer two limitations for mapping: bias in spatial coverage and imprecision in spatial location. This article presents an approach to minimizing the impacts of these two limitations of citizen data using geospatial analysis techniques. The approach reduces location imprecision by adopting a frequency-sampling strategy to identify representative presence locations from areas over which citizens observed the geographic phenomenon. The approach compensates for the spatial bias by weighting presence locations with cumulative visibility (the frequency at which a given location can be seen by local citizens). As a case study to demonstrate the principle, this approach was applied to map the habitat suitability of the black-and-white snub-nosed monkey (Rhinopithecus bieti) in Yunnan, China. Sightings of R. bieti were elicited from local citizens using a geovisualization platform and then processed with the proposed approach to predict a habitat suitability map. Presence locations of R. bieti recorded by biologists through intensive field tracking were used to validate the predicted habitat suitability map. Validation showed that the continuous Boyce index (Bcont(0.1)) calculated on the suitability map was 0.873 (95% CI: [0.810, 0.917]), indicating that the map was highly consistent with the field-observed distribution of R. bieti. Bcont(0.1) was much lower (0.173) for the suitability map predicted based on citizen data when location imprecision was not reduced and even lower (?0.048) when there was no compensation for spatial bias. This indicates that the proposed approach effectively minimized the impacts of location imprecision and spatial bias in citizen data and therefore effectively improved the quality of mapped spatial variation using citizen data. It further implies that, with the application of geospatial analysis techniques to properly account for limitations in citizen data, valuable information embedded in such data can be extracted and used for scientific mapping.  相似文献   

3.
Location-allocation modeling is an important area of research in spatial optimization and GIScience. A large number of analytical models for location-allocation analysis have been developed in the past 50 years to meet the requirements of different planning and spatial-analytic applications, ranging from the location of emergency response units (EMS) to warehouses and transportation hubs. Despite their great number, many location-allocation models are intrinsically linked to one another. A well-known example is the theoretical link between the classic p-median problem and coverage location problems. Recently, Lei and Church showed that a large number of classic and new location models can be posed as special case problems of a new modeling construct called the vector assignment ordered median problem (VAOMP). Lei and Church also reported extremely high computational complexity in optimally solving the best integer linear programming (ILP) formulation developed for the VAOMP even for medium-sized problems in certain cases.

In this article, we develop an efficient unified solver for location-allocation analysis based on the VAOMP model without using ILP solvers. Our aim is to develop a fast heuristic algorithm based on the Tabu Search (TS) meta-heuristic, and message passing interface (MPI) suitable for obtaining optimal or near-optimal solutions for the VAOMP in a real-time environment. The unified approach is particularly interesting from the perspective of GIScience and spatial decision support systems (DSS) as it makes it possible to solve a wide variety of location models in a unified manner in a GIS environment. Computational results show that the TS method can often obtain in seconds, solutions that are better than those obtained using the ILP-based approach in hours or a day.  相似文献   

4.
Effects of spatial autocorrelation (SAC), or spatial structure, have often been neglected in the conventional models of pedogeomorphological processes. Based on soil, vegetation, and topographic data collected in a coastal dunefield in western Korea, this research developed three soil moisture–landscape models, each incorporating SAC at fine, broad, and multiple scales, respectively, into a non-spatial ordinary least squares (OLS) model. All of these spatially explicit models showed better performance than the OLS model, as consistently indicated by R2, Akaike’s information criterion, and Moran’s I. In particular, the best model was proved to be the one using spatial eigenvector mapping, a technique that accounts for spatial structure at multiple scales simultaneously. After including SAC, predictor variables with greater inherent spatial structure underwent more reduction in their predictive power than those with less structure. This finding implies that the environmental variables pedogeomorphologists have perceived important in the conventional regression modeling may have a reduced predictive power in reality, in cases where they possess a significant amount of SAC. This research demonstrates that accounting for spatial structure not only helps to avoid the violation of statistical assumptions, but also allows a better understanding of dynamic soil hydrological processes occurring at different spatial scales.  相似文献   

5.
空间数据挖掘的地理案例推理方法及试验   总被引:2,自引:0,他引:2  
杜云艳  温伟  曹锋 《地理研究》2009,28(5):1285-1296
从空间数据挖掘的角度谈地理案例推理方法,认为地理案例推理是面向问题的一种空间数据挖掘方法。针对这一思想进行了基于地理案例的空间数据挖掘具体算法介绍。首先在明确地理案例具体定义的基础上,给出了面向问题的空间数据挖掘地理案例界定和组织方法;其次,鉴于地理空间的自然地带性和区域分异性规律的影响,深入探讨了地理案例自身或其间所可能存在的相互依赖和相互制约关系,并给出了采用粗糙集方法进行地理案例内蕴空间关系的定量挖掘方法;第三,针对地理案例表达时考虑的空间特征和空间关系的不同,给出了三种状况下的空间相似性计算模型;最后,以土地利用这一典型的地学现象为例,给出具体实例,一方面进行土地利用问题的定量分析与推测;另一方面,通过实例展示地理案例推理方法在地学问题求解以及空间数据定量分析上的特点和优势。  相似文献   

6.
With the ubiquity of advanced web technologies and location-sensing hand held devices, citizens regardless of their knowledge or expertise, are able to produce spatial information. This phenomenon is known as volunteered geographic information (VGI). During the past decade VGI has been used as a data source supporting a wide range of services, such as environmental monitoring, events reporting, human movement analysis, disaster management, etc. However, these volunteer-contributed data also come with varying quality. Reasons for this are: data is produced by heterogeneous contributors, using various technologies and tools, having different level of details and precision, serving heterogeneous purposes, and a lack of gatekeepers. Crowd-sourcing, social, and geographic approaches have been proposed and later followed to develop appropriate methods to assess the quality measures and indicators of VGI. In this article, we review various quality measures and indicators for selected types of VGI and existing quality assessment methods. As an outcome, the article presents a classification of VGI with current methods utilized to assess the quality of selected types of VGI. Through these findings, we introduce data mining as an additional approach for quality handling in VGI.  相似文献   

7.
Seabed sediment textural parameters such as mud, sand and gravel content can be useful surrogates for predicting patterns of benthic biodiversity. Multibeam sonar mapping can provide near-complete spatial coverage of high-resolution bathymetry and backscatter data that are useful in predicting sediment parameters. Multibeam acoustic data collected across a ~1000 km2 area of the Carnarvon Shelf, Western Australia, were used in a predictive modelling approach to map eight seabed sediment parameters. Four machine learning models were used for the predictive modelling: boosted decision tree, random forest decision tree, support vector machine and generalised regression neural network. The results indicate overall satisfactory statistical performance, especially for %Mud, %Sand, Sorting, Skewness and Mean Grain Size. The study also demonstrates that predictive modelling using the combination of machine learning models has provided the ability to generate prediction uncertainty maps. However, the single models were shown to have overall better prediction performance than the combined models. Another important finding was that choosing an appropriate set of explanatory variables, through a manual feature selection process, was a critical step for optimising model performance. In addition, machine learning models were able to identify important explanatory variables, which are useful in identifying underlying environmental processes and checking predictions against the existing knowledge of the study area. The sediment prediction maps obtained in this study provide reliable coverage of key physical variables that will be incorporated into the analysis of covariance of physical and biological data for this area.  相似文献   

8.
In this article, we have proposed a simple diagnostic statistical procedure for testing the order of dependence of a spatial process. The proposed test is of nonparametric nature, and it is able to deal with potential nonlinear spatial dependencies. An added value is that from a methodological point of view, the new test is based on symbolic dynamics and hence on information theory. We characterized the behavior of a symbolic entropy measure in the presence of spatial dependencies of order higher than 1. The good power performance of the new method in detecting higher order spatial lags is notable and gives rise to an expectation that it may form a suitable basis for constructive specification searches.  相似文献   

9.
Categorical spatial data, such as land use classes and socioeconomic statistics data, are important data sources in geographical information science (GIS). The investigation of spatial patterns implied in these data can benefit many aspects of GIS research, such as classification of spatial data, spatial data mining, and spatial uncertainty modeling. However, the discrete nature of categorical data limits the application of traditional kriging methods widely used in Gaussian random fields. In this article, we present a new probabilistic method for modeling the posterior probability of class occurrence at any target location in space-given known class labels at source data locations within a neighborhood around that prediction location. In the proposed method, transition probabilities rather than indicator covariances or variograms are used as measures of spatial structure and the conditional or posterior (multi-point) probability is approximated by a weighted combination of preposterior (two-point) transition probabilities, while accounting for spatial interdependencies often ignored by existing approaches. In addition, the connections of the proposed method with probabilistic graphical models (Bayesian networks) and weights of evidence method are also discussed. The advantages of this new proposed approach are analyzed and highlighted through a case study involving the generation of spatial patterns via sequential indicator simulation.  相似文献   

10.
Abstract

Mapping by sampling and prediction of local and regional values of two-dimensional surfaces is a frequent, complex task in geographical information systems. This article describes a method for the approximation of two-dimensional surfaces by optimizing sample size, arrangement and prediction accuracy simultaneously. First, a grid of an ancillary data set is approximated by a quadtree to determine a predefined number of homogeneous mapping units. This approximation is optimal in the sense of minimizing Kullback-divergence between the quadtree and the grid of ancillary data. Then, samples are taken from each mapping unit. The performance of this sampling has been tested against other sampling strategies (regular and random) and found to be superior in reconstructing the grid using three interpolation techniques (inverse squared Euclidean distance, kriging, and Thiessen-polygonization). Finally, the discrepancy between the ancillary grid and the surface to be mapped is modelled by different levels and spatial structures of noise. Conceptually this method is advantageous in cases when sampling strata cannot be well defined a priori and the spatial structure of the phenomenon to be mapped is not known, but ancillary information (e.g., remotely-sensed data), corresponding to its spatial pattern, is available.  相似文献   

11.
During the last two decades, a variety of models have been applied to understand and predict changes in land use. These models assign a single-attribute label to each spatial unit at any particular time of the simulation. This is not realistic because mixed use of land is quite common. A more detailed classification allowing the modelling of mixed land use would be desirable for better understanding and interpreting the evolution of the use of land. A possible solution is the multi-label (ML) concept where each spatial unit can belong to multiple classes simultaneously. For example, a cluster of summer houses at a lake in a forested area should be classified as water, forest and residential (built-up). The ML concept was introduced recently, and it belongs to the machine learning field. In this article, the ML concept is introduced and applied in land-use modelling. As a novelty, we present a land-use change model that allows ML class assignment using the k nearest neighbour (kNN) method that derives a functional relationship between land use and a set of explanatory variables. A case study with a rich data-set from Luxembourg using biophysical data from aerial photography is described. The model achieves promising results based on the well-known ML evaluation criteria. The application described in this article highlights the value of the multi-label k nearest neighbour method (MLkNN) for land-use modelling.  相似文献   

12.
ABSTRACT

This paper proposes a new classification method for spatial data by adjusting prior class probabilities according to local spatial patterns. First, the proposed method uses a classical statistical classifier to model training data. Second, the prior class probabilities are estimated according to the local spatial pattern and the classifier for each unseen object is adapted using the estimated prior probability. Finally, each unseen object is classified using its adapted classifier. Because the new method can be coupled with both generative and discriminant statistical classifiers, it performs generally more accurately than other methods for a variety of different spatial datasets. Experimental results show that this method has a lower prediction error than statistical classifiers that take no spatial information into account. Moreover, in the experiments, the new method also outperforms spatial auto-logistic regression and Markov random field-based methods when an appropriate estimate of local prior class distribution is used.  相似文献   

13.
Most of the literature to date proposes approximations to the determinant of a positive definite × n spatial covariance matrix (the Jacobian term) for Gaussian spatial autoregressive models that fail to support the analysis of massive georeferenced data sets. This paper briefly surveys this literature, recalls and refines much simpler Jacobian approximations, presents selected eigenvalue estimation techniques, summarizes validation results (for estimated eigenvalues, Jacobian approximations, and estimation of a spatial autocorrelation parameter), and illustrates the estimation of the spatial autocorrelation parameter in a spatial autoregressive model specification for cases as large as n = 37,214,101. The principal contribution of this paper is to the implementation of spatial autoregressive model specifications for any size of georeferenced data set. Its specific additions to the literature include (1) new, more efficient estimation algorithms; (2) an approximation of the Jacobian term for remotely sensed data forming incomplete rectangular regions; (3) issues of inference; and (4) timing results.  相似文献   

14.
Abstract

Recent developments in theory and computer software mean that it is now relatively straightforward to evaluate how attribute errors are propagated through quantitative spatial models in GIS. A major problem, however, is to estimate the errors associated with the inputs to these spatial models. A first approach is to use the root mean square error, but in many cases it is better to estimate the errors from the degree of spatial variation and the method used for mapping. It is essential to decide at an early stage whether one should use a discrete model of spatial variation (DMSV—homogeneous areas, abrupt boundaries), a continuous model (CMSV—a continuously varying regionalized variable field) or a mixture of both (MMSV—mixed model of spatial variation). Maps of predictions and prediction error standard deviations are different in all three cases, and it is crucial for error estimation which model of spatial variation is used. The choice of model has been insufficiently studied in depth, but can be based on prior information about the kinds of spatial processes and patterns that are present, or on validation results. When undetermined it is sensible to adopt the MMSV in order to bypass the rigidity of the DMSV and CMSV. These issues are explored and illustrated using data on the mean highest groundwater level in a polder area in the Netherlands.  相似文献   

15.
ABSTRACT

The efficiency of public investments and services has been of interest to geographic researchers for several decades. While in the private sector inefficiency often leads to higher prices, loss of competitiveness, and loss of business, in the public sector inefficiency in service provision does not necessarily lead to immediate changes. In many cases, it is not an entirely easy task to analyze a particular service as appropriate data may be difficult to obtain and hidden in detailed budgets. In this paper, we develop an integrative approach that uses cyber search, Geographic Information System (GIS), and spatial optimization to estimate the spatial efficiency of fire protection services in Los Angeles (LA) County. We develop a cyber-search process to identify current deployment patterns of fire stations across the major urban region of LA County. We compare the results of our search to existing databases. Using spatial optimization, we estimate the level of deployment that is needed to meet desired coverage levels based upon the location of an ideal fire station pattern, and then compare this ideal level of deployment to the existing system as a means of estimating spatial efficiency. GIS is adopted throughout the paper to simulate the demand locations, to conduct location-based spatial analysis, to visualize fire station data, and to map model simulation results. Finally, we show that the existing system in LA County has considerable room for improvement. The methodology presented in this paper is both novel and groundbreaking, and the automated assessments are readily transferable to other counties and jurisdictions.  相似文献   

16.
Geo-tagged travel photos on social networks often contain location data such as points of interest (POIs), and also users’ travel preferences. In this paper, we propose a hybrid ensemble learning method, BAyes-Knn, that predicts personalized tourist routes for travelers by mining their geographical preferences from these location-tagged data. Our method trains two types of base classifiers to jointly predict the next travel destination: (1) The K-nearest neighbor (KNN) classifier quantifies users’ location history, weather condition, temperature and seasonality and uses a feature-weighted distance model to predict a user’s personalized interests in an unvisited location. (2) A Bayes classifier introduces a smooth kernel function to estimate a-priori probabilities of features and then combines these probabilities to predict a user’s latent interests in a location. All the outcomes from these subclassifiers are merged into one final prediction result by using the Borda count voting method. We evaluated our method on geo-tagged Flickr photos and Beijing weather data collected from 1 January 2005 to 1 July 2016. The results demonstrated that our ensemble approach outperformed 12 other baseline models. In addition, the results showed that our framework has better prediction accuracy than do context-aware significant travel-sequence-patterns recommendations and frequent travel-sequence patterns.  相似文献   

17.
In the field of digital terrain analysis (DTA), the principle and method of uncertainty in surface area calculation (SAC) have not been deeply developed and need to be further studied. This paper considers the uncertainty of data sources from the digital elevation model (DEM) and SAC in DTA to perform the following investigations: (a) truncation error (TE) modeling and analysis, (b) modeling and analysis of SAC propagation error (PE) by using Monte-Carlo simulation techniques and spatial autocorrelation error to simulate DEM uncertainty. The simulation experiments show that (a) without the introduction of the DEM error, higher DEM resolution and lower terrain complexity lead to smaller TE and absolute error (AE); (b) with the introduction of the DEM error, the DEM resolution and terrain complexity influence the AE and standard deviation (SD) of the SAC, but the trends by which the two values change may be not consistent; and (c) the spatial distribution of the introduced random error determines the size and degree of the deviation between the calculated result and the true value of the surface area. This study provides insights regarding the principle and method of uncertainty in SACs in geographic information science (GIScience) and provides guidance to quantify SAC uncertainty.  相似文献   

18.
史文娇  张沫 《地理学报》2022,77(11):2890-2901
土壤粒径(砂粒、粉粒和黏粒)是各种陆表过程和生态系统服务评估等模型的关键参数。作为一种土壤成分数据,土壤粒径的空间预测方法有和为1(或100%)等特殊要求,其空间分布精度受预测方法影响较大。本文针对土壤粒径相较于其他土壤属性的特殊性,提出了土壤粒径空间预测方法框架,综述了土壤粒径数据变换、空间插值和精度验证等系列方法,总结了提升土壤粒径空间预测精度的各种途径,包括通过有效的数据变换改善数据分布、结合数据分布特点选择合适的预测方法、结合辅助变量提升制图精度和分布合理性、使用混合模型提升插值精度、使用多成分联合模拟模型提升预测的系统性等。最后,提出了今后土壤粒径空间预测方法研究的未来方向,包括从考虑数据变换原理和机制角度改善数据分布、发展多成分联合模拟模型和高精度曲面建模方法,以及引入土壤粒径函数曲线并与随机模拟结合等。  相似文献   

19.
20.
空间关联规则挖掘研究进展   总被引:7,自引:0,他引:7  
随着空间数据获取技术的进步, 空间数据量日益增大, 已超出人们的分析能力。传统的空 间数据分析方法只能进行简单的数据分析, 无法满足人们获取知识的需要。空间关联规则是空间 数据挖掘一个基本的任务, 是从具有海量、多维、多尺度、不确定性边界等特性的空间数据中进行 知识发现的重要方法。本文从基本概念、分类、挖掘过程、挖掘方法、目前研究成果等方面对其进 行综述, 重点阐述了空间关联规则挖掘效率的改进策略、基于不确定空间信息的挖掘方法、挖掘 过程及结果的可视化、弱空间关联规则的挖掘方法等。通过对现有空间关联规则研究成果和存在 问题的深入剖析, 指出了其未来主要的发展方向。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号