首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Rendering large volumes of vector data is computationally intensive and therefore time consuming, leading to lower efficiency and poorer interactive experience. Graphics processing units (GPUs) are powerful tools in data parallel processing but lie idle most of the time. In this study, we propose an approach to improve the performance of vector data rendering by using the parallel computing capability of many‐core GPUs. Vertex transformation, largely a mathematical calculation that does not require communication with the host storage device, is a time‐consuming procedure because all coordinates of each vector feature need to be transformed to screen vertices. Use of a GPU enables optimization of a general‐purpose mathematical calculation, enabling the procedure to be executed in parallel on a many‐core GPU and optimized effectively. This study mainly focuses on: (1) an organization and storage strategy for vector data based on equal pitch alignment, which can adapt to the GPU's calculating characteristics; (2) a paging‐coalescing transfer and memory access strategy for vector data between the CPU and the GPU; and (3) a balancing allocation strategy to take full advantage of all processing cores of the GPU. Experimental results demonstrate that the approach proposed can significantly improve the efficiency of vector data rendering.  相似文献   

2.
随着对地立体观测体系的建立,遥感大数据不断累积。传统基于文件、景/幅式的影像组织方式,时空基准不够统一,集中式存储不利于大规模并行分析。对地观测大数据分析仍缺乏一套统一的数据模型与基础设施理论。近年来,数据立方体的研究为对地观测领域大数据分析基础设施提供了前景。基于统一的分析就绪型多维数据模型和集成对地观测数据分析功能,可构建一个基于数据立方的对地观测大数据分析基础设施。因此,本文提出了一个面向大规模分析的多源对地观测时空立方体,相较于现有的数据立方体方法,强调多源数据的统一组织、基于云计算的立方体处理模式以及基于人工智能优化的立方体计算。研究有助于构建时空大数据分析的新框架,同时建立与商业智能领域的数据立方体关联,为时空大数据建立统一的时空组织模型,支持大范围、长时序的快速大规模对地观测数据分析。本文在性能上与开源数据立方做了对比,结果证明提出的多源对地观测时空立方体在处理性能上具有明显优势。  相似文献   

3.
Abstract

The geospatial sciences face grand information technology (IT) challenges in the twenty-first century: data intensity, computing intensity, concurrent access intensity and spatiotemporal intensity. These challenges require the readiness of a computing infrastructure that can: (1) better support discovery, access and utilization of data and data processing so as to relieve scientists and engineers of IT tasks and focus on scientific discoveries; (2) provide real-time IT resources to enable real-time applications, such as emergency response; (3) deal with access spikes; and (4) provide more reliable and scalable service for massive numbers of concurrent users to advance public knowledge. The emergence of cloud computing provides a potential solution with an elastic, on-demand computing platform to integrate – observation systems, parameter extracting algorithms, phenomena simulations, analytical visualization and decision support, and to provide social impact and user feedback – the essential elements of the geospatial sciences. We discuss the utilization of cloud computing to support the intensities of geospatial sciences by reporting from our investigations on how cloud computing could enable the geospatial sciences and how spatiotemporal principles, the kernel of the geospatial sciences, could be utilized to ensure the benefits of cloud computing. Four research examples are presented to analyze how to: (1) search, access and utilize geospatial data; (2) configure computing infrastructure to enable the computability of intensive simulation models; (3) disseminate and utilize research results for massive numbers of concurrent users; and (4) adopt spatiotemporal principles to support spatiotemporal intensive applications. The paper concludes with a discussion of opportunities and challenges for spatial cloud computing (SCC).  相似文献   

4.
三维多视角立体视觉算法(patch-based multi-view stereo,PMVS)以其良好的三维重建效果广泛应用于数字城市等领域,但用于大规模计算时算法的执行效率低下。针对此,提出了一种细粒度并行优化方法,从任务划分和负载均衡、主系统存储和GPU存储、通信开销等3方面加以优化;同时,设计了基于面片的PMVS算法特征提取的GPU和多线程并行改造方法,实现了CPUs_GPUs多粒度协同并行。实验结果表明,基于CPU多线程策略能实现4倍加速比,基于统一计算设备架构(compute unified device architecture,CUDA)并行策略能实现最高34倍加速比,而提出的策略在CUDA并行策略的基础上实现了30%的性能提升,可以用于其他领域大数据处理中快速调度计算资源。  相似文献   

5.
Abstract

Global challenges (such as economy and natural hazards) and technology advancements have triggered international leaders and organizations to rethink geosciences and Digital Earth in the new decade. The next generation visions pose grand challenges for infrastructure, especially computing infrastructure. The gradual establishment of cloud computing as a primary infrastructure provides new capabilities to meet the challenges. This paper reviews research conducted using cloud computing to address geoscience and Digital Earth needs within the context of an integrated Earth system. We also introduce the five papers selected through a rigorous review process as exemplar research in using cloud capabilities to address the challenges. The literature and research demonstrate that spatial cloud computing provides unprecedented new capabilities to enable Digital Earth and geosciences in the twenty-first century in several aspects: (1) virtually unlimited computing power for addressing big data storage, sharing, processing, and knowledge discovering challenges, (2) elastic, flexible, and easy-to-use computing infrastructure to facilitate the building of the next generation geospatial cyberinfrastructure, CyberGIS, CloudGIS, and Digital Earth, (3) seamless integration environment that enables mashing up observation, data, models, problems, and citizens, (4) research opportunities triggered by global challenges that may lead to breakthroughs in relevant fields including infrastructure building, GIScience, computer science, and geosciences, and (5) collaboration supported by cloud computing and across science domains, agencies, countries to collectively address global challenges from policy, management, system engineering, acquisition, and operation aspects.  相似文献   

6.
A geospatial cyberinfrastructure is needed to support advanced GIScience research and education activities. However, the heterogeneous and distributed nature of geospatial resources creates enormous obstacles for building a unified and interoperable geospatial cyberinfrastructure. In this paper, we propose the Geospatial Service Web (GSW) to underpin the development of a future geospatial cyberinfrastructure. The GSW excels over the traditional spatial data infrastructure by providing a highly intelligent geospatial middleware to integrate various geospatial resources through the Internet based on interoperable Web service technologies. The development of the GSW focuses on the establishment of a platform where data, information, and knowledge can be shared and exchanged in an interoperable manner. Theoretically, we describe the conceptual framework and research challenges for GSW, and then introduce our recent research toward building a GSW. A research agenda for building a GSW is also presented in the paper.  相似文献   

7.
Geospatial Agents, Agents Everywhere . . .   总被引:1,自引:0,他引:1  
The use of the related terms “agent‐based”, “multi‐agent”, “software agent” and “intelligent agent” have witnessed significant growth in the Geographic Information Science (GIScience) literature in the past decade. These terms usually refer to both artificial life agents that simulate human and animal behavior and software agents that support human‐computer interactions. In this article we first comprehensively review both types of agents. Then we argue that both these categories of agents borrow from Artificial Intelligence (AI) research, requiring them to share the characteristics of and be similar to AI agents. We also argue that geospatial agents form a distinct category of AI agents because they are explicit about geography and geographic data models. Our overall goal is to first capture the diversity of, and then define and categorize GIScience agent research into geospatial agents, thereby capturing the diversity of agent‐oriented architectures and applications that have been developed in the recent past to present a holistic review of geospatial agents.  相似文献   

8.
Today, many real‐time geospatial applications (e.g. navigation and location‐based services) involve data‐ and/or compute‐intensive geoprocessing tasks where performance is of great importance. Cloud computing, a promising platform with a large pool of storage and computing resources, could be a practical solution for hosting vast amounts of data and for real‐time processing. In this article, we explored the feasibility of using Google App Engine (GAE), the cloud computing technology by Google, for a module in navigation services, called Integrated GNSS (iGNSS) QoS prediction. The objective of this module is to predict quality of iGNSS positioning solutions for prospective routes in advance. iGNSS QoS prediction involves the real‐time computation of large Triangulated Irregular Networks (TINs) generated from LiDAR data. We experimented with the Google App Engine (GAE) and stored a large TIN for two geoprocessing operations (proximity and bounding box) required for iGNSS QoS prediction. The experimental results revealed that while cloud computing can potentially be used for development and deployment of data‐ and/or compute‐intensive geospatial applications, current cloud platforms require improvements and special tools for handling real‐time geoprocessing, such as iGNSS QoS prediction, efficiently. The article also provides a set of general guidelines for future development of real‐time geoprocessing in clouds.  相似文献   

9.
Disaster response operations require fast and coordinated actions based on real‐time disaster situation information. Although crowdsourced geospatial data applications have been demonstrated to be valuable tools for gathering real‐time disaster situation information, they only provide limited utility for disaster response coordination because of the lack of semantic compatibility and interoperability. To help overcome the semantic incompatibility and heterogeneity problems, we use Geospatial Semantic Web (GSW) technologies. We then combine GSW technologies with Web Feature Service requests to access multiple servers. However, a GSW‐based geographic information system often has poor performance due to the complex geometric computations required. The objective of this research is to explore how to use optimization techniques to improve the performance of an interoperable geographic situation‐awareness system (IGSAS) based on GSW technologies for disaster response. We conducted experiments to evaluate various client‐side optimization techniques for improving the performance of an IGSAS prototype for flood disaster response in New Haven, Connecticut. Our experimental results show that the developed prototype can greatly reduce the runtime costs of geospatial semantic queries through on‐the‐fly spatial indexing, tile‐based rendering, efficient algorithms for spatial join, and caching, especially for those spatial‐join geospatial queries that involve a large number of spatial features and heavy geometric computation.  相似文献   

10.
Semantic standardization is an integral part of sharing data for GIS and spatial analysis. It is part of a broader rubric of interoperability or the ability to share geographic information across multiple platforms and contexts. GIScience researchers have made considerable progress towards understanding and addressing the multiple challenges involved in achieving interoperability. For local government agencies interested in sharing spatial data, however, current interoperability approaches based on object-oriented data models represent idealistic solutions to problems of semantic heterogeneity that often exceed the level of sophistication and funding available. They are waiting for the market to decide how interoperability should be resolved. In order to assist in this transition, this paper presents a rule-based Visual Basic application to standardize the semantics of simple spatial entities using several classification systems. We use the example of well-log data, and argue that this approach enables agencies to share and structure data effectively in an interim period during which market and research standards for semantic interoperability are being determined. It contributes to a geospatial data infrastructure, while allowing agencies to share spatial data in a manner consistent with their level of expertise and existing data structures.  相似文献   

11.
A spatial web portal (SWP) provides a web-based gateway to discover, access, manage, and integrate worldwide geospatial resources through the Internet and has the access characteristics of regional to global interest and spiking. Although various technologies have been adopted to improve SWP performance, enabling high-speed resource access for global users to better support Digital Earth remains challenging because of the computing and communication intensities in the SWP operation and the dynamic distribution of end users. This paper proposes a cloud-enabled framework for high-speed SWP access by leveraging elastic resource pooling, dynamic workload balancing, and global deployment. Experimental results demonstrate that the new SWP framework outperforms the traditional computing infrastructure and better supports users of a global system such as Digital Earth. Reported methodologies and framework can be adopted to support operational geospatial systems, such as monitoring national geographic state and spanning across regional and global geographic extent.  相似文献   

12.
Big geospatial data is an emerging sub‐area of geographic information science, big data, and cyberinfrastructure. Big geospatial data poses two unique challenges. First, raster and vector data structures and analyses have developed on largely separate paths for the last 20 years. This is creating an impediment to geospatial researchers seeking to utilize big data platforms that do not promote heterogeneous data types. Second, big spatial data repositories have yet to be integrated with big data computation platforms in ways that allow researchers to spatio‐temporally analyze big geospatial datasets. IPUMS‐Terra, a National Science Foundation cyberInfrastructure project, addresses these challenges by providing a unified framework of integrated geospatial services which access, analyze, and transform big heterogeneous spatio‐temporal data. As IPUMS‐Terra's data volume grows, we seek to integrate geospatial platforms that will scale geospatial analyses and address current bottlenecks within our system. However, our work shows that there are still unresolved challenges for big geospatial analysis. The most pertinent is that there is a lack of a unified framework for conducting scalable integrated vector and raster data analysis. We conducted a comparative analysis between PostgreSQL with PostGIS and SciDB and concluded that SciDB is the superior platform for scalable raster zonal analyses.  相似文献   

13.
王良清 《现代测绘》2006,29(6):39-41
省级地理空间数据交换中心是国家空间信息基础设施建设的核心内容之一。本文针对现阶段在我国建设省级地理空间数据交换中心的问题,分析了其基本建设内容,提出了需解决和应用的关键技术。  相似文献   

14.
利用集群架构和分布式并行可视化工具 VisIt,编写了自定义插件,实现了基于大规模地球系统格网组织下的全球科学数据并行可视化,并设计实验对其并行可视化性能进行了对比分析.实验发现: VisIt完成一次渲染的加速比及并行效率随着核数的增加逐渐降低;采用 GPU 渲染,可以很好地提高并行渲染的效率.但在核数和 GPU 个数同步增加的情况下,由于核间通信、 GPU 间通信以及核- GPU 间通信等,VisIt一次渲染的并行运行时间并无明显降低.随着数据量增加,VisIt对单位数据量的运行时间却逐渐减低.实验表明,VisIt可较高效地完成大数据量的并行渲染.该方法和结论可供地学领域大规模海量数据可视化研究参考.  相似文献   

15.
Due to high data volume, massive spatial data requires considerable computing power for real‐time processing. Currently, high performance clusters are the only economically viable solution given the development of multicore technology and computer component cost reduction in recent years. Massive spatial data processing demands heavy I/O operations, however, and should be characterized as a data‐intensive application. Data‐intensive application parallelization strategies, such as decomposition, scheduling and load‐balance, are much different from that of traditional compute‐intensive applications. In this article we introduce a Split‐and‐Merge paradigm for spatial data processing and also propose a robust parallel framework in a cluster environment to support this paradigm. The Split‐and‐Merge paradigm efficiently exploits data parallelism for massive data processing. The proposed framework is based on the open‐source TORQUE project and hosted on a multicore‐enabled Linux cluster. A specific data‐aware scheduling algorithm was designed to exploit data sharing between tasks and decrease the data communication time. Two LiDAR point cloud algorithms, IDW interpolation and Delaunay triangulation, were implemented on the proposed framework to evaluate its efficiency and scalability. Experimental results demonstrate that the system provides efficient performance speedup.  相似文献   

16.
当前云计算的发展已能支持高性能的地理空间服务,比如在数字城市和电子商务等行业。Apache基金支持下的开源软件框架Hadoop,可以用来构建一个云环境的集群用来存储和处理高性能的地理空间数据。开放地理空间联盟(OGC)的Web三维服务(W3DS)就是这样一个很好的三维的地理空间数据服务标准。在标准的云计算环境下将是一个更好的应用示范。基于此,本文研究了OGC的W3DS服务在云计算环境下的实验结果。实验采用Apache的Hadoop框架作为三维地理空间信息服务实验展示的基础。实验结果对展示高性能的三维地理空间信息提供了有价值的参考。  相似文献   

17.
GeoPW: Laying Blocks for the Geospatial Processing Web   总被引:2,自引:0,他引:2  
Recent advances in Web‐related technologies have significantly promoted the wide sharing and integrated analysis of distributed geospatial data. Geospatial applications often involve diverse sources of data and complex geoprocessing functions. Existing Web‐based GIS focuses more on access to distributed geospatial data. In scientific problem solving, the ability to carry out geospatial analysis is essential to geoscientific discovery. This article presents the design and implementation of GeoPW, a set of services providing geoprocessing functions over the Web. The concept of the Geospatial Processing Web is discussed to address the geoprocessing demands in the emerging information infrastructure, and the role of GeoPW in establishing the Geospatial Processing Web is identified. The services in GeoPW are implemented by developing middleware that wraps legacy GIS analysis components to provide a large number of geoprocessing utilities over the Web. These services are open and accessible to the public, and they support integrated geoprocessing on the Web.  相似文献   

18.
If sites, cities, and landscapes are captured at different points in time using technology such as LiDAR, large collections of 3D point clouds result. Their efficient storage, processing, analysis, and presentation constitute a challenging task because of limited computation, memory, and time resources. In this work, we present an approach to detect changes in massive 3D point clouds based on an out‐of‐core spatial data structure that is designed to store data acquired at different points in time and to efficiently attribute 3D points with distance information. Based on this data structure, we present and evaluate different processing schemes optimized for performing the calculation on the CPU and GPU. In addition, we present a point‐based rendering technique adapted for attributed 3D point clouds, to enable effective out‐of‐core real‐time visualization of the computation results. Our approach enables conclusions to be drawn about temporal changes in large highly accurate 3D geodata sets of a captured area at reasonable preprocessing and rendering times. We evaluate our approach with two data sets from different points in time for the urban area of a city, describe its characteristics, and report on applications.  相似文献   

19.
Quantitative remote sensing retrieval algorithms help understanding the dynamic aspects of Digital Earth. However, the Big Data and complex models in Digital Earth pose grand challenges for computation infrastructures. In this article, taking the aerosol optical depth (AOD) retrieval as a study case, we exploit parallel computing methods for high efficient geophysical parameter retrieval. We present an efficient geocomputation workflow for the AOD calculation from the Moderate Resolution Imaging Spectroradiometer (MODIS) satellite data. According to their individual potential for parallelization, several procedures were adapted and implemented for a successful parallel execution on multi-core processors and Graphics Processing Units (GPUs). The benchmarks in this paper validate the high parallel performance of the retrieval workflow with speedups of up to 5.x on a multi-core processor with 8 threads and 43.x on a GPU. To specifically address the time-consuming model retrieval part, hybrid parallel patterns which combine the multi-core processor’s and the GPU’s compute power were implemented with static and dynamic workload distributions and evaluated on two systems with different CPU–GPU configurations. It is shown that only the dynamic hybrid implementation leads to a greatly enhanced overall exploitation of the heterogeneous hardware environment in varying circumstances.  相似文献   

20.
Geospatial Semantic Web promises better retrieval geospatial information for Digital Earth systems by explicitly representing the semantics of data through ontologies. It also promotes sharing and reuse of geospatial data by encoding it in Semantic Web languages, such as RDF, to form geospatial knowledge base. For many applications, rapid retrieval of spatial data from the knowledge base is critical. However, spatial data retrieval using the standard Semantic Web query language – Geo-SPARQL – can be very inefficient because the data in the knowledge base are no longer indexed to support efficient spatial queries. While recent research has been devoted to improving query performance on general knowledge base, it is still challenging to support efficient query of the spatial data with complex topological relationships. This research introduces a query strategy to improve the query performance of geospatial knowledge base by creating spatial indexing on-the-fly to prune the search space for spatial queries and by parallelizing the spatial join computations within the queries. We focus on improving the performance of Geo-SPARQL queries on knowledge bases encoded in RDF. Our initial experiments show that the proposed strategy can greatly reduce the runtime costs of Geo-SPARQL query through on-the-fly spatial indexing and parallel execution.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号