全文获取类型
收费全文 | 5356篇 |
免费 | 593篇 |
国内免费 | 743篇 |
专业分类
测绘学 | 2526篇 |
大气科学 | 799篇 |
地球物理 | 884篇 |
地质学 | 927篇 |
海洋学 | 572篇 |
天文学 | 148篇 |
综合类 | 582篇 |
自然地理 | 254篇 |
出版年
2024年 | 11篇 |
2023年 | 54篇 |
2022年 | 159篇 |
2021年 | 215篇 |
2020年 | 241篇 |
2019年 | 304篇 |
2018年 | 188篇 |
2017年 | 303篇 |
2016年 | 269篇 |
2015年 | 317篇 |
2014年 | 341篇 |
2013年 | 426篇 |
2012年 | 414篇 |
2011年 | 400篇 |
2010年 | 243篇 |
2009年 | 307篇 |
2008年 | 288篇 |
2007年 | 344篇 |
2006年 | 306篇 |
2005年 | 251篇 |
2004年 | 197篇 |
2003年 | 184篇 |
2002年 | 123篇 |
2001年 | 112篇 |
2000年 | 80篇 |
1999年 | 79篇 |
1998年 | 104篇 |
1997年 | 82篇 |
1996年 | 65篇 |
1995年 | 54篇 |
1994年 | 42篇 |
1993年 | 37篇 |
1992年 | 29篇 |
1991年 | 24篇 |
1990年 | 21篇 |
1989年 | 24篇 |
1988年 | 14篇 |
1987年 | 12篇 |
1986年 | 7篇 |
1985年 | 3篇 |
1984年 | 7篇 |
1983年 | 1篇 |
1982年 | 2篇 |
1981年 | 1篇 |
1980年 | 2篇 |
1979年 | 1篇 |
1976年 | 1篇 |
1954年 | 3篇 |
排序方式: 共有6692条查询结果,搜索用时 93 毫秒
81.
粗差发现和定位能力与相关系数的关系 总被引:2,自引:0,他引:2
根据粗差判断方程中的判断矩阵和两个统计检验量之间相关系数的函数式,论证了两种不同的研究方法所确定的粗差不能定位的数学模型实际上是相等的。通过算例,不仅说明两种研究方法对观测量不能定位粗差的判断是一致的,而且使用判断矩阵研究观测量的粗差发现和定位能力会更加方便简单。 相似文献
82.
平差系统的模型误差及其识别方法研究 总被引:2,自引:0,他引:2
论述了模型误差影响参数估值的一些理论问题,指出了随机模型误差和函数模型误差之间的相互作用和转化。为讨论平差系统最优模型的选取,给出了与现有文献将模型误差纳入平差系统的思路不同的一个估计和识别模型误差的理论基础公式,由此导出了相应的实用公式,给出了平差系统模型的优选方法。 相似文献
83.
介绍了一个建筑物阴影检测的模型。首先利用摄影测量学原理来计算阴影坐标。即用数字表面模型(digitalsurfacemodel,缩写为DSM)和太阳高度和方位来计算建筑物阴影的空间坐标,并由相机模型计算出每个阴影单元对应的扫描行和相机空间坐标。由高度场光线跟踪判断阴影的可见性,对可见阴影计算出它在投影图像上的坐标。然后在这个结果的基础上再对图像进行阴影的细分割。 相似文献
84.
85.
本文探讨了附加系统参数平差法在带有测距系统误差的导线网中的应用,说明了当系统误差比较显著时.采用附加系统参数平差法的可行性、必要性及其优越性。 相似文献
86.
Uncertainties in polar motion and length-of-day measurements are evaluated empirically using several data series from the space-geodetic techniques of the global positioning system (GPS), satellite laser ranging (SLR), and very long baseline interferometry (VLBI) during 1997–2002. In the evaluation procedure employed here, known as the three-corner hat (TCH) technique, the signal common to each series is eliminated by forming pair-wise differences between the series, thus requiring no assumed values for the “truth” signal. From the variances of the differenced series, the uncertainty of each series can be recovered when reasonable assumptions are made about the correlations between the series. In order to form the pair-wise differences, the series data must be given at the same epoch. All measurement data sets studied here were sampled at noon (UTC); except for the VLBI series, whose data are interpolated to noon and whose UT1 values are also numerically differentiated to obtain LOD. The numerical error introduced to the VLBI values by the interpolation and differentiation is shown to be comparable in magnitude to the values determined by the TCH method for the uncertainties of the VLBI series. The TCH estimates for the VLBI series are corrupted by such numerical errors mostly as a result of the relatively large data intervals. Of the remaining data sets studied here, it is found that the IGS Final combined series has the smallest polar motion and length-of-day uncertainties. 相似文献
87.
Some theory problems affecting parameter estimation are discussed in this paper. Influence and transformation between errors
of stochastic and functional models is pointed out as well. For choosing the best adjustment model, a formula, which is different
from the literatures existing methods, for estimating and identifying the model error, is proposed. On the basis of the proposed
formula, an effective approach of selecting the best model of adjustment system is given.
Project supported by the Open Research Fund Program of the Key Laboratory of Geospace Environment and Geodesy, Ministry of
Education, Wuhan University (No. 905276031-04-10). 相似文献
88.
M. Kern T. Preimesberger M. Allesch R. Pail J. Bouman R. Koop 《Journal of Geodesy》2005,78(9):509-519
The satellite missions CHAMP, GRACE, and GOCE mark the beginning of a new era in gravity field determination and modeling. They provide unique models of the global stationary gravity field and its variation in time. Due to inevitable measurement errors, sophisticated pre-processing steps have to be applied before further use of the satellite measurements. In the framework of the GOCE mission, this includes outlier detection, absolute calibration and validation of the SGG (satellite gravity gradiometry) measurements, and removal of temporal effects. In general, outliers are defined as observations that appear to be inconsistent with the remainder of the data set. One goal is to evaluate the effect of additive, innovative and bulk outliers on the estimates of the spherical harmonic coefficients. It can be shown that even a small number of undetected outliers (<0.2 of all data points) can have an adverse effect on the coefficient estimates. Consequently, concepts for the identification and removal of outliers have to be developed. Novel outlier detection algorithms are derived and statistical methods are presented that may be used for this purpose. The methods aim at high outlier identification rates as well as small failure rates. A combined algorithm, based on wavelets and a statistical method, shows best performance with an identification rate of about 99%. To further reduce the influence of undetected outliers, an outlier detection algorithm is implemented inside the gravity field solver (the Quick-Look Gravity Field Analysis tool was used). This results in spherical harmonic coefficient estimates that are of similar quality to those obtained without outliers in the input data. 相似文献
89.
Investigating the propagation mechanism of unmodelled systematic errors on coordinate time series estimated using least squares 总被引:6,自引:8,他引:6
The propagation of unmodelled systematic errors into coordinate time series computed using least squares is investigated,
to improve the understanding of unexplained signals and apparent noise in geodetic (especially GPS) coordinate time series.
Such coordinate time series are invariably based on a functional model linearised using only zero and first-order terms of
a (Taylor) series expansion about the approximate coordinates of the unknown point. The effect of such truncation errors is
investigated through the derivation of a generalised systematic error model for the simple case of range observations from
a single known reference point to a point which is assumed to be at rest by the least squares model but is in fact in motion.
The systematic error function for a one pseudo-satellite two-dimensional case, designed to be as simple but as analogous to
GPS positioning as possible, is quantified. It is shown that the combination of a moving reference point and unmodelled periodic
displacement at the unknown point of interest, due to ocean tide loading, for example, results in an output coordinate time
series containing many periodic terms when only zero and first-order expansion terms are used in the linearisation of the
functional model. The amplitude, phase and period of these terms is dependent on the input amplitude, the locations of the
unknown point and reference point, and the period of the reference point's motion. The dominant output signals that arise
due to truncation errors match those found in coordinate time series obtained from both simulated data and real three-dimensional
GPS data. 相似文献
90.
The purpose of this article is to describe the development of a remotely sensed, historical land-cover change database for the northwestern quarter of Chihuahua, Mexico, The database consists of multi-temporal land-cover classifications and change detection images. The database is developed to facilitate future investigations that examine urban–rural linkages as possible drivers of rural land-use and land-cover changes. To develop the needed land-cover change database, this study uses the North American Landsat Characterization (NALC) MSS triplicates because of their temporal depth and spatial breadth. Challenges exist, however, to effective classification and change detection using the NALC triplicates, including illumination differences across multiple scenes and periods caused by topographic and solar variations and the lack of ground reference data for historic periods. Therefore, creation of the database is a four step process. First, extensive pre-processing is performed to enhance comparability of multi-date images. Pre-processing includes topographic correction, mosaic creation and multi-date radiance normalization. Second, ancillary sources of land-cover data are combined with visual interpretations of enhanced images to define reference pixels used to classify the images using the maximum likelihood algorithm. Third, classification accuracy is assessed. Fourth, post-classification change detection is performed. Results indicate significant image improvements after pre-processing that permit very good overall classification (86.26% classified correctly) and change detection. To conclude, findings are presented that indicate significant changes to arid grasslands/shrublands and forest resources in mountainous regions. 相似文献