首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
The purpose of object matching in conflation is to identify corresponding objects in different data sets that represent the same real-world entity. This article presents an improved linear object matching approach, named the optimization and iterative logistic regression matching (OILRM) method, which combines the optimization model and logistic regression model to obtain a better matching result by detecting incorrect matches and missed matches that are included in the result obtained from the optimization (Opt) method for object matching in conflation. The implementation of the proposed OILRM method was demonstrated in a comprehensive case study of Shanghai, China. The experimental results showed the following. (1) The Opt method can determine most of the optimal one-to-one matching pairs under the condition of minimizing the total distance of all matching pairs without setting empirical thresholds. However, the matching accuracy and recall need to be further improved. (2) The proposed OILRM method can detect incorrect matches and missed matches and resolve the issues of one-to-many and many-to-many matching relationships with a higher matching recall. (3) In the case where the source data sets become more complicated, the matching accuracy and recall based on the proposed OILRM method are much better than those based on the Opt method.  相似文献   

2.
Recently, researchers have introduced deep learning methods such as convolutional neural networks (CNN) to model spatio-temporal data and achieved better results than those with conventional methods. However, these CNN-based models employ a grid map to represent spatial data, which is unsuitable for road-network-based data. To address this problem, we propose a deep spatio-temporal residual neural network for road-network-based data modeling (DSTR-RNet). The proposed model constructs locally-connected neural network layers (LCNR) to model road network topology and integrates residual learning to model the spatio-temporal dependency. We test the DSTR-RNet by predicting the traffic flow of Didi cab service, in an 8-km2 region with 2,616 road segments in Chengdu, China. The results demonstrate that the DSTR-RNet maintains the spatial precision and topology of the road network as well as improves the prediction accuracy. We discuss the prediction errors and compare the prediction results to those of grid-based CNN models. We also explore the sensitivity of the model to its parameters; this will aid the application of this model to network-based data modeling.  相似文献   

3.
4.
Different versions of the Web Coverage Service (WCS) schemas of the Open Geospatial Consortium (OGC) reflect semantic conflict. When applying the extended FRAG-BASE schema-matching approach (a schema-matching method based on COMA++, including an improved schema decomposition algorithm and schema fragments identification algorithm, which enable COMA++-based support to OGC Web Service schema matching), the average recall of WCS schema matching is only 72%, average precision is only 82% and average overall is only 57%. To improve the quality of multi-version WCS retrieval, we propose a schema-matching method that measures node semantic similarity (NSS). The proposed method is based on WordNet, conjunctive normal form and a vector space model. A hybrid algorithm based on label meanings and annotations is designed to calculate the similarity between label concepts. We translate the semantic relationships between nodes into a propositional formula and verify the validity of this formula to confirm the semantic relationships. The algorithm first computes the label and node concepts and then calculates the conceptual relationship between the labels. Finally, the conceptual relationship between nodes is computed. We then use the NSS method in experiments on different versions of WCS. Results show that the average recall of WCS schema matching is greater than 83%; average precision reaches 92%; and average overall is 67%.  相似文献   

5.
ABSTRACT

The increasing popularity of Location-Based Social Networks (LBSNs) and the semantic enrichment of mobility data in several contexts in the last years has led to the generation of large volumes of trajectory data. In contrast to GPS-based trajectories, LBSN and context-aware trajectories are more complex data, having several semantic textual dimensions besides space and time, which may reveal interesting mobility patterns. For instance, people may visit different places or perform different activities depending on the weather conditions. These new semantically rich data, known as multiple-aspect trajectories, pose new challenges in trajectory classification, which is the problem that we address in this paper. Existing methods for trajectory classification cannot deal with the complexity of heterogeneous data dimensions or the sequential aspect that characterizes movement. In this paper we propose MARC, an approach based on attribute embedding and Recurrent Neural Networks (RNNs) for classifying multiple-aspect trajectories, that tackles all trajectory properties: space, time, semantics, and sequence. We highlight that MARC exhibits good performance especially when trajectories are described by several textual/categorical attributes. Experiments performed over four publicly available datasets considering the Trajectory-User Linking (TUL) problem show that MARC outperformed all competitors, with respect to accuracy, precision, recall, and F1-score.  相似文献   

6.
Map matching method is a fundamental preprocessing technique for massive probe vehicle data. Various transportation applications need map matching methods to provide highly accurate and stable results. However, most current map matching approaches employ elementary geometric or topological measures, which may not be sufficient to encode the characteristic of realistic driving paths, leading to inefficiency and inaccuracy, especially in complex road networks. To address these issues, this article presents a novel map matching method, based on the measure of curvedness of Global Positioning System (GPS) trajectories. The curvature integral, which measures the curvedness feature of GPS trajectories, is considered to be one of the major matching characteristics that constrain pairwise matching between the two adjacent GPS track points. In this article, we propose the definition of the curvature integral in the context of map matching, and develop a novel accurate map matching algorithm based on the curvedness feature. Using real-world probe vehicles data, we show that the curvedness feature (CURF) constrained map matching method outperforms two classical methods for accuracy and stability under complicated road environments.  相似文献   

7.
ABSTRACT

Urban landmarks are of significant importance to spatial cognition and route navigation. However, the current landmark extraction methods mainly focus on the visual salience of landmarks and are insufficient for obtaining high extraction accuracy when the size of the geographical dataset varies. This study introduces a random forests (RF) classifier combining with the synthetic minority oversampling technique (SMOTE) in urban landmark extraction. Both GIS and social sensing data are employed to quantify the structural and cognitive salience of the examined urban features, which are available from basic spatial databases or mainstream web service application programming interfaces (APIs). The results show that the SMOTE-RF model performs well in urban landmark extraction, with the values of recall, precision, F-measure and AUC reaching 0.851, 0.831, 0.841 and 0.841, respectively. Additionally, this method is suitable for both large and small geographical datasets. The ranking of variable importance given by this model further indicates that certain cognitive measures – such as feature class, Weibo popularity and Bing popularity – can serve as crucial factors for determining a landmark. The optimal variable combination for landmark extraction is also acquired, which might provide support for eliminating the variable selection requirement in other landmark extraction methods.  相似文献   

8.
Monitoring and predicting traffic conditions are of utmost importance in reacting to emergency events in time and for computing the real-time shortest travel-time path. Mobile sensors, such as GPS devices and smartphones, are useful for monitoring urban traffic due to their large coverage area and ease of deployment. Many researchers have employed such sensed data to model and predict traffic conditions. To do so, we first have to address the problem of associating GPS trajectories with the road network in a robust manner. Existing methods rely on point-by-point matching to map individual GPS points to a road segment. However, GPS data is imprecise due to noise in GPS signals. GPS coordinates can have errors of several meters and, therefore, direct mapping of individual points is error prone. Acknowledging that every GPS point is potentially noisy, we propose a radically different approach to overcome inaccuracy in GPS data. Instead of focusing on a point-by-point approach, our proposed method considers the set of relevant GPS points in a trajectory that can be mapped together to a road segment. This clustering approach gives us a macroscopic view of the GPS trajectories even under very noisy conditions. Our method clusters points based on the direction of movement as a spatial-linear cluster, ranks the possible route segments in the graph for each group, and searches for the best combination of segments as the overall path for the given set of GPS points. Through extensive experiments on both synthetic and real datasets, we demonstrate that, even with highly noisy GPS measurements, our proposed algorithm outperforms state-of-the-art methods in terms of both accuracy and computational cost.  相似文献   

9.
In integration of road maps modeled as road vector data, the main task is matching pairs of objects that represent, in different maps, the same segment of a real-world road. In an ad hoc integration, the matching is done for a specific need and, thus, is performed in real time, where only a limited preprocessing is possible. Usually, ad hoc integration is performed as part of some interaction with a user and, hence, the matching algorithm is required to complete its task in time that is short enough for human users to provide feedback to the application, that is, in no more than a few seconds. Such interaction is typical of services on the World Wide Web and to applications in car-navigation systems or in handheld devices.

Several algorithms were proposed in the past for matching road vector data; however, these algorithms are not efficient enough for ad hoc integration. This article presents algorithms for ad hoc integration of maps in which roads are represented as polylines. The main novelty of these algorithms is in using only the locations of the endpoints of the polylines rather than trying to match whole lines. The efficiency of the algorithms is shown both analytically and experimentally. In particular, these algorithms do not require the existence of a spatial index, and they are more efficient than an alternative approach based on using a grid index. Extensive experiments using various maps of three different cities show that our approach to matching road networks is efficient and accurate (i.e., it provides high recall and precision).

General Terms:Algorithms, Experimentation  相似文献   

10.
等高线蕴含的历史高程信息可有效延长地形研究的时间序列,有利于深入挖掘地形变化长期规律,然而,图幅接边处的高程属性错误降低了等高线的数据质量,制约着等高线高程信息的实际应用。针对这一问题,该文提出一种基于层次格网索引的图幅接边处等高线高程错误识别和自动修正方法:首先,将层次格网索引与方向性二邻域算法相结合,以减少数据重复计算;然后,利用等高线空间位置标签及快速排序算法构建强空间位置关系,解决图幅接边处等高线匹配的准确性问题;最后,以高程冲突位点为驱动因子进行逻辑判断,实现等高线高程错误的识别及自动修正。实验结果表明:该方法运算效率较未进行效率优化时提高了203倍,接边处等高线高程错误识别与修正精度的最大值分别达97.71%和91.40%;相较于现有方法,该方法在精度和效率方面表现更佳,对区域性错误和变形等高线具有更高的适用性。  相似文献   

11.
ABSTRACT

Spatial interpolation is a traditional geostatistical operation that aims at predicting the attribute values of unobserved locations given a sample of data defined on point supports. However, the continuity and heterogeneity underlying spatial data are too complex to be approximated by classic statistical models. Deep learning models, especially the idea of conditional generative adversarial networks (CGANs), provide us with a perspective for formalizing spatial interpolation as a conditional generative task. In this article, we design a novel deep learning architecture named conditional encoder-decoder generative adversarial neural networks (CEDGANs) for spatial interpolation, therein combining the encoder-decoder structure with adversarial learning to capture deep representations of sampled spatial data and their interactions with local structural patterns. A case study on elevations in China demonstrates the ability of our model to achieve outstanding interpolation results compared to benchmark methods. Further experiments uncover the learned spatial knowledge in the model’s hidden layers and test the potential to generalize our adversarial interpolation idea across domains. This work is an endeavor to investigate deep spatial knowledge using artificial intelligence. The proposed model can benefit practical scenarios and enlighten future research in various geographical applications related to spatial prediction.  相似文献   

12.
ABSTRACT

Defining and identifying duplicate records in a dataset is a challenging task which grows more complex when the modeled entities themselves are hard to delineate. In the geospatial domain, it may not be clear where a mountain, stream, or valley ends and begins, a problem carried over when such entities are catalogued in gazetteers. In this paper, we take two gazetteers, GeoNames and SwissNames3D, and perform matching – identifying records in each that are about the same entity – across a sample of natural feature records. We first perform rule-based matching, establishing competitive results, then apply machine learning using Random Forests, a method well-suited to the matching task. We report on the performance of a wider array of matching features than has been previously studied, including domain-specific ones such as feature type, land cover class, and elevation. Our results show an increase in performance using machine learning over rules, with a notable performance gain from considering feature types, but negligible gains from other specialized matching features. We argue that future work in this area should strive to be more reproducible and report results on a realistic testing pipeline including candidate selection, feature extraction, and classification.  相似文献   

13.
Toponym matching, i.e. pairing strings that represent the same real-world location, is a fundamental problemfor several practical applications. The current state-of-the-art relies on string similarity metrics, either specifically developed for matching place names or integrated within methods that combine multiple metrics. However, these methods all rely on common sub-strings in order to establish similarity, and they do not effectively capture the character replacements involved in toponym changes due to transliterations or to changes in language and culture over time. In this article, we present a novel matching approach, leveraging a deep neural network to classify pairs of toponyms as either matching or nonmatching. The proposed network architecture uses recurrent nodes to build representations from the sequences of bytes that correspond to the strings that are to be matched. These representations are then combined and passed to feed-forward nodes, finally leading to a classification decision. We present the results of a wide-ranging evaluation on the performance of the proposed method, using a large dataset collected from the GeoNames gazetteer. These results show that the proposed method can significantly outperform individual similarity metrics from previous studies, as well as previous methods based on supervised machine learning for combining multiple metrics.  相似文献   

14.
ABSTRACT

Short-term traffic forecasting on large street networks is significant in transportation and urban management, such as real-time route guidance and congestion alleviation. Nevertheless, it is very challenging to obtain high prediction accuracy with reasonable computational cost due to the complex spatial dependency on the traffic network and the time-varying traffic patterns. To address these issues, this paper develops a residual graph convolution long short-term memory (RGC-LSTM) model for spatial-temporal data forecasting considering the network topology. This model integrates a new graph convolution operator for spatial modelling on networks and a residual LSTM structure for temporal modelling considering multiple periodicities. The proposed model has few parameters, low computational complexity, and a fast convergence rate. The framework is evaluated on both the 10-min traffic speed data from Shanghai, China and the 5-min Caltrans Performance Measurement System (PeMS) traffic flow data. Experiments show the advantages of the proposed approach over various state-of-the-art baselines, as well as consistent performance across different datasets.  相似文献   

15.
利用电感耦合等离子体质谱(ICP-MS)技术,对比分析了直接稀释法和基体匹配法测定含盐水样中微量铀的精密度、准确度和回收率。结果表明,两种方法的铀标准曲线线性相关系数均大于0.999 9,精密度、准确度和回收率都满足样品测定要求,二者均可用于盐湖水中微量铀的测定。但基体匹配法稳定性高,重现性好,更适合于盐湖水中微量铀的快速、准确测定。  相似文献   

16.
Recently, points of interest (POIs) recommendation has evolved into a hot research topic with real-world applications. In this paper, we propose a novel semantics-enhanced density-based clustering algorithm SEM-DTBJ-Cluster, to extract semantic POIs from GPS trajectories. We then take into account three different factors (popularity, temporal and geographical features) that can influence the recommendation score of a POI. We characterize the impacts caused by popularity, temporal and geographical information, by using different scoring functions based on three developed recommendation models. Finally, we combine the three scoring functions together and obtain a unified framework PTG-Recommend for recommending candidate POIs for a mobile user. To the best of our knowledge, this work is the first that considers popularity, temporal and geographical information together. Experimental results on two real-world data sets strongly demonstrate that our framework is robust and effective, and outperforms the baseline recommendation methods in terms of precision and recall.  相似文献   

17.
Geospatial data matching is an important prerequisite for data integration, change detection and data updating. At present, crowdsourcing geospatial data are attracting considerable attention with its significant potential for timely and cost-effective updating of geospatial data and Geographical Information Science (GIS) applications. To integrate the available and up-to-date information of multi-source geospatial data, this article proposes a heuristic probabilistic relaxation road network matching method. The proposed method starts with an initial probabilistic matrix according to the dissimilarities in the shapes and then integrates the relative compatibility coefficient of neighbouring candidate pairs to iteratively update the initial probabilistic matrix until the probabilistic matrix is globally consistent. Finally, the initial 1:1 matching pairs are selected on the basis of probabilities that are calculated and refined on the basis of the structural similarity of the selected matching pairs. A process of matching is then implemented to find M:N matching pairs. Matching between OpenStreetMap network data and professional road network data shows that our method is independent of matching direction, successfully matches 1:0 (Null), 1:1 and M:N pairs, and achieves a robust matching precision of above 95%.  相似文献   

18.
基于GIS的二维非结构化剖分网格优化   总被引:2,自引:1,他引:1  
非结构化网格剖分是数值模拟的关键技术之一,网格质量直接影响到计算的收敛性和精确度。在GIS辅助建立非结构化网格空间拓扑关系的基础上,针对GIS和实际研究问题给出Spring-Laplace方法——一种新的单元尺度函数定义,在GIS空间算法下利用该方法优化节点位置,并基于推进阵面算法的思想,结合空间邻近拓扑关系实现了三角剖分节点和网格的重新编号算法,方便了开边界条件的赋值,提高了计算效率。实例表明,该方法大大提高了网格生成质量,能适应FVCOM数值模型对非结构化网格剖分的要求,其收敛速度快,具有较高的运算效率。  相似文献   

19.
With the increase in the number of applications using digital vector maps and the development of surveying techniques, a large volume of GIS (geographic information system) vector maps having high accuracy and precision is being produced. However, to achieve their effective transmission while preserving their high positional quality, these large amounts of vector map data need to be compressed. This paper presents a compression method based on a bin space partitioning data structure, which preserves a high-level accuracy and exact precision of spatial data. To achieve this, the proposed method a priori divides a map into rectangular local regions and classifies the bits of each object in the local regions to three types of bins, defined as category bin (CB), direction bin (DB), and accuracy bin (AB). Then, it encodes objects progressively using the properties of the classified bins, such as adjacency and orientation, to obtain the optimum compression ratio. Experimental results verify that our method can encode vector map data constituting less than 20% of the original map data at a 1-cm accuracy degree and that constituting less than 9% at a 1-m accuracy degree. In addition, its compression efficiency is greater than that of previous methods, whereas its complexity is lower for close to real-time applications.  相似文献   

20.
随着我国地膜使用面积的增加和人们对土壤微塑料污染问题的日益关注,大尺度的地膜遥感识别已成为农业生产管理、土壤污染防治的必要手段。针对地膜光谱反射特征的复杂性以及基于单一遥感影像光谱特征识别方法错分率高等问题,该文以河北省邯郸市邱县为试验区,利用GF-1数据的空间细节与Sentinel-2数据的光谱信息进行NN Diffuse Pan Sharpening融合,据此建立地膜识别的特征矩阵(NDVI、MNDWI、NDBI、IBI、PSI),基于该特征矩阵可实现自动阈值地膜分层分类识别。多种方法的地膜识别结果精度对比表明:多源光学遥感数据融合方法的总体精度为94.87%,Kappa系数达0.89,显著优于基于单一数据源的深度学习法的精度(93.14%)以及基于传统机器学习分类方法的支持向量机(85.91%)和随机森林分类法(86.78%)的精度;通过与Sentinel-2多光谱影像融合,弥补了GF-1数据光谱分辨率低的缺陷,实现了多源数据在地膜识别中的优势互补,可为相关部门农业规划与管理以及生态环境保护等研究提供大尺度、高精度的地膜分布参考数据。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号