首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
We present new insights on the time-averaged surface velocities, convergence and extension rates along arc-normal transects in Kumaon, Garhwal and Kashmir–Himachal regions in the Indian Himalaya from 13 years of high-precision Global Positioning System (GPS) time series (1995–2008) derived from GPS data at 14 GPS permanent and 42 campaign stations between $29.5{-}35^{\circ }\hbox {N}$ and $76{-}81^{\circ }\hbox {E}$ . The GPS surface horizontal velocities vary significantly from the Higher to Lesser Himalaya and are of the order of 30 to 48 mm/year NE in ITRF 2005 reference frame, and 17 to 2 mm/year SW in an India fixed reference frame indicating that this region is accommodating less than 2 cm/year of the India–Eurasia plate motion ( ${\sim }4~\hbox {cm/year}$ ). The total arc-normal shortening varies between ${\sim }10{-}14~\hbox {mm/year}$ along the different transects of the northwest Himalayan wedge, between the Indo-Tsangpo suture to the north and the Indo-Gangetic foreland to the south indicating high strain accumulation in the Himalayan wedge. This convergence is being accommodated differentially along the arc-normal transects; ${\sim } 5{-}10~\hbox {mm/year}$ in Lesser Himalaya and 3–4 mm/year in Higher Himalaya south of South Tibetan Detachment. Most of the convergence in the Lesser Himalaya of Garhwal and Kumaon is being accommodated just south of the Main Central Thrust fault trace, indicating high strain accumulation in this region which is also consistent with the high seismic activity in this region. In addition, for the first time an arc-normal extension of ${\sim }6~\hbox {mm/year}$ has also been observed in the Tethyan Himalaya of Kumaon. Inverse modeling of GPS-derived surface deformation rates in Garhwal and Kumaon Himalaya using a single dislocation indicate that the Main Himalayan Thrust is locked from the surface to a depth of ${\sim }15{-}20~\hbox {km}$ over a width of 110 km with associated slip rate of ${\sim }16{-}18~\hbox {mm/year}$ . These results indicate that the arc-normal rates in the Northwest Himalaya have a complex deformation pattern involving both convergence and extension, and rigorous seismo-tectonic models in the Himalaya are necessary to account for this pattern. In addition, the results also gave an estimate of co-seismic and post-seismic motion associated with the 1999 Chamoli earthquake, which is modeled to derive the slip and geometry of the rupture plane.  相似文献   

2.
Error analysis of the NGS’ surface gravity database   总被引:1,自引:1,他引:0  
Are the National Geodetic Survey’s surface gravity data sufficient for supporting the computation of a 1 cm-accurate geoid? This paper attempts to answer this question by deriving a few measures of accuracy for this data and estimating their effects on the US geoid. We use a data set which comprises ${\sim }1.4$ million gravity observations collected in 1,489 surveys. Comparisons to GRACE-derived gravity and geoid are made to estimate the long-wavelength errors. Crossover analysis and $K$ -nearest neighbor predictions are used for estimating local gravity biases and high-frequency gravity errors, and the corresponding geoid biases and high-frequency geoid errors are evaluated. Results indicate that 244 of all 1,489 surface gravity surveys have significant biases ${>}2$  mGal, with geoid implications that reach 20 cm. Some of the biased surveys are large enough in horizontal extent to be reliably corrected by satellite-derived gravity models, but many others are not. In addition, the results suggest that the data are contaminated by high-frequency errors with an RMS of ${\sim }2.2$  mGal. This causes high-frequency geoid errors of a few centimeters in and to the west of the Rocky Mountains and in the Appalachians and a few millimeters or less everywhere else. Finally, long-wavelength ( ${>}3^{\circ }$ ) surface gravity errors on the sub-mGal level but with large horizontal extent are found. All of the south and southeast of the USA is biased by +0.3 to +0.8 mGal and the Rocky Mountains by $-0.1$ to $-0.3$  mGal. These small but extensive gravity errors lead to long-wavelength geoid errors that reach 60 cm in the interior of the USA.  相似文献   

3.
Well credited and widely used ionospheric models, such as the International Reference Ionosphere or NeQuick, describe the variation of the electron density with height by means of a piecewise profile tied to the F2-peak parameters: the electron density, $N_m \mathrm{F2}$ N m F 2 , and the height, $h_m \mathrm{F2}$ h m F 2 . Accurate values of these parameters are crucial for retrieving reliable electron density estimations from those models. When direct measurements of these parameters are not available, the models compute the parameters using the so-called ITU-R database, which was established in the early 1960s. This paper presents a technique aimed at routinely updating the ITU-R database using radio occultation electron density profiles derived from GPS measurements gathered from low Earth orbit satellites. Before being used, these radio occultation profiles are validated by fitting to them an electron density model. A re-weighted Least Squares algorithm is used for down-weighting unreliable measurements (occasionally, entire profiles) and to retrieve $N_m \mathrm{F2}$ N m F 2 and $h_m \mathrm{F2}$ h m F 2 values—together with their error estimates—from the profiles. These values are used to monthly update the database, which consists of two sets of ITU-R-like coefficients that could easily be implemented in the IRI or NeQuick models. The technique was tested with radio occultation electron density profiles that are delivered to the community by the COSMIC/FORMOSAT-3 mission team. Tests were performed for solstices and equinoxes seasons in high and low-solar activity conditions. The global mean error of the resulting maps—estimated by the Least Squares technique—is between $0.5\times 10^{10}$ 0.5 × 10 10 and $3.6\times 10^{10}$ 3.6 × 10 10 elec/m $^{-3}$ ? 3 for the F2-peak electron density (which is equivalent to 7 % of the value of the estimated parameter) and from 2.0 to 5.6 km for the height ( $\sim $ 2 %).  相似文献   

4.
We can map zenith wet delays onto precipitable water with a conversion factor, but in order to calculate the exact conversion factor, we must precisely calculate its key variable $T_\mathrm{m}$ . Yao et al. (J Geod 86:1125–1135, 2012. doi:10.1007/s00190-012-0568-1) established the first generation of global $T_\mathrm{m}$ model (GTm-I) with ground-based radiosonde data, but due to the lack of radiosonde data at sea, the model appears to be abnormal in some areas. Given that sea surface temperature varies less than that on land, and the GPT model and the Bevis $T_\mathrm{m}$ $T_\mathrm{s}$ relationship are accurate enough to describe the surface temperature and $T_\mathrm{m}$ , this paper capitalizes on the GPT model and the Bevis $T_\mathrm{m}$ $T_\mathrm{s}$ relationship to provide simulated $T_\mathrm{m}$ at sea, as a compensation for the lack of data. Combined with the $T_\mathrm{m}$ from radiosonde data, we recalculated the GTm model coefficients. The results show that this method not only improves the accuracy of the GTm model significantly at sea but also improves that on land, making the GTm model more stable and practically applicable.  相似文献   

5.
We develop a slope correction model to improve the accuracy of mean sea surface topography models as well as marine gravity models. The correction is greatest above ocean trenches and large seamounts where the slope of the geoid exceeds 100  \(\upmu \) rad. In extreme cases, the correction to the mean sea surface height is 40 mm and the correction to the along-track altimeter slope is 1–2  \(\upmu \) rad which maps into a 1–2 mGal gravity error. Both corrections are easily applied using existing grids of sea surface slope from satellite altimetry.  相似文献   

6.
The LLL algorithm, introduced by Lenstra et al. (Math Ann 261:515–534, 1982), plays a key role in many fields of applied mathematics. In particular, it is used as an effective numerical tool for preconditioning the integer least-squares problems arising in high-precision geodetic positioning and Global Navigation Satellite Systems (GNSS). In 1992, Teunissen developed a method for solving these nearest-lattice point (NLP) problems. This method is referred to as Lambda (for Least-squares AMBiguity Decorrelation Adjustment). The preconditioning stage of Lambda corresponds to its decorrelation algorithm. From an epistemological point of view, the latter was devised through an innovative statistical approach completely independent of the LLL algorithm. Recent papers pointed out some similarities between the LLL algorithm and the Lambda-decorrelation algorithm. We try to clarify this point in the paper. We first introduce a parameter measuring the orthogonality defect of the integer basis in which the NLP problem is solved, the LLL-reduced basis of the LLL algorithm, or the $\Lambda $ -basis of the Lambda method. With regard to this problem, the potential qualities of these bases can then be compared. The $\Lambda $ -basis is built by working at the level of the variance-covariance matrix of the float solution, while the LLL-reduced basis is built by working at the level of its inverse. As a general rule, the orthogonality defect of the $\Lambda $ -basis is greater than that of the corresponding LLL-reduced basis; these bases are however very close to one another. To specify this tight relationship, we present a method that provides the dual LLL-reduced basis of a given $\Lambda $ -basis. As a consequence of this basic link, all the recent developments made on the LLL algorithm can be applied to the Lambda-decorrelation algorithm. This point is illustrated in a concrete manner: we present a parallel $\Lambda $ -type decorrelation algorithm derived from the parallel LLL algorithm of Luo and Qiao (Proceedings of the fourth international C $^*$ conference on computer science and software engineering. ACM Int Conf P Series. ACM Press, pp 93–101, 2012).  相似文献   

7.
This research represents a continuation of the investigation carried out in the paper of Petrovskaya and Vershkov (J Geod 84(3):165–178, 2010) where conventional spherical harmonic series are constructed for arbitrary order derivatives of the Earth gravitational potential in the terrestrial reference frame. The problem of converting the potential derivatives of the first and second orders into geopotential models is studied. Two kinds of basic equations for solving this problem are derived. The equations of the first kind represent new non-singular non-orthogonal series for the geopotential derivatives, which are constructed by means of transforming the intermediate expressions for these derivatives from the above-mentioned paper. In contrast to the spherical harmonic expansions, these alternative series directly depend on the geopotential coefficients ${\bar{{C}}_{n,m}}$ and ${\bar{{S}}_{n,m}}$ . Each term of the series for the first-order derivatives is represented by a sum of these coefficients, which are multiplied by linear combinations of at most two spherical harmonics. For the second-order derivatives, the geopotential coefficients are multiplied by linear combinations of at most three spherical harmonics. As compared to existing non-singular expressions for the geopotential derivatives, the new expressions have a more simple structure. They depend only on the conventional spherical harmonics and do not depend on the first- and second-order derivatives of the associated Legendre functions. The basic equations of the second kind are inferred from the linear equations, constructed in the cited paper, which express the coefficients of the spherical harmonic series for the first- and second-order derivatives in terms of the geopotential coefficients. These equations are converted into recurrent relations from which the coefficients ${\bar{{C}}_{n,m}}$ and ${\bar{{S}}_{n,m}}$ are determined on the basis of the spherical harmonic coefficients of each derivative. The latter coefficients can be estimated from the values of the geopotential derivatives by the quadrature formulas or the least-squares approach. The new expressions of two kinds can be applied for spherical harmonic synthesis and analysis. In particular, they might be incorporated in geopotential modeling on the basis of the orbit data from the CHAMP, GRACE and GOCE missions, and the gradiometry data from the GOCE mission.  相似文献   

8.
Fast error analysis of continuous GNSS observations with missing data   总被引:3,自引:0,他引:3  
One of the most widely used method for the time-series analysis of continuous Global Navigation Satellite System (GNSS) observations is Maximum Likelihood Estimation (MLE) which in most implementations requires $\mathcal{O }(n^3)$ operations for $n$ observations. Previous research by the authors has shown that this amount of operations can be reduced to $\mathcal{O }(n^2)$ for observations without missing data. In the current research we present a reformulation of the equations that preserves this low amount of operations, even in the common situation of having some missing data.Our reformulation assumes that the noise is stationary to ensure a Toeplitz covariance matrix. However, most GNSS time-series exhibit power-law noise which is weakly non-stationary. To overcome this problem, we present a Toeplitz covariance matrix that provides an approximation for power-law noise that is accurate for most GNSS time-series.Numerical results are given for a set of synthetic data and a set of International GNSS Service (IGS) stations, demonstrating a reduction in computation time of a factor of 10–100 compared to the standard MLE method, depending on the length of the time-series and the amount of missing data.  相似文献   

9.
M-estimation with probabilistic models of geodetic observations   总被引:1,自引:1,他引:0  
The paper concerns \(M\) -estimation with probabilistic models of geodetic observations that is called \(M_{\mathcal {P}}\) estimation. The special attention is paid to \(M_{\mathcal {P}}\) estimation that includes the asymmetry and the excess kurtosis, which are basic anomalies of empiric distributions of errors of geodetic or astrometric observations (in comparison to the Gaussian errors). It is assumed that the influence function of \(M_{\mathcal {P}}\) estimation is equal to the differential equation that defines the system of the Pearson distributions. The central moments \(\mu _{k},\, k=2,3,4\) , are the parameters of that system and thus, they are also the parameters of the chosen influence function. The \(M_{\mathcal {P}}\) estimation that includes the Pearson type IV and VII distributions ( \(M_{\mathrm{PD(l)}}\) method) is analyzed in great detail from a theoretical point of view as well as by applying numerical tests. The chosen distributions are leptokurtic with asymmetry which refers to the general characteristic of empirical distributions. Considering \(M\) -estimation with probabilistic models, the Gram–Charlier series are also applied to approximate the models in question ( \(M_{\mathrm{G-C}}\) method). The paper shows that \(M_{\mathcal {P}}\) estimation with the application of probabilistic models belongs to the class of robust estimations; \(M_{\mathrm{PD(l)}}\) method is especially effective in that case. It is suggested that even in the absence of significant anomalies the method in question should be regarded as robust against gross errors while its robustness is controlled by the pseudo-kurtosis.  相似文献   

10.
Non-linear station motions in epoch and multi-year reference frames   总被引:5,自引:5,他引:0  
In the conventions of the International Earth Rotation and Reference Systems Service (e.g. IERS Conventions 2010), it is recommended that the instantaneous station position, which is fixed to the Earth’s crust, is described by a regularized station position and conventional correction models. Current realizations of the International Terrestrial Reference Frame use a station position at a reference epoch and a constant velocity to describe the motion of the regularized station position in time. An advantage of this parameterization is the possibility to provide station coordinates of high accuracy over a long time span. Various publications have shown that residual non-linear station motions can reach a magnitude of a few centimeters due to not considered loading effects. Consistently estimated parameters like the Earth Orientation Parameters (EOP) may be affected if these non-linear station motions are neglected. In this paper, we investigate a new approach, which is based on a frequent (e.g. weekly) estimation of station positions and EOP from a combination of epoch normal equations of the space geodetic techniques Global Positioning System (GPS), Satellite Laser Ranging (SLR) and Very Long Baseline Interferometry (VLBI). The resulting time series of epoch reference frames are studied in detail and are compared with the conventional secular approach. It is shown that both approaches have specific advantages and disadvantages, which are discussed in the paper. A major advantage of the frequently estimated epoch reference frames is that the non-linear station motions are implicitly taken into account, which is a major limiting factor for the accuracy of the secular frames. Various test computations and comparisons between the epoch and secular approach are performed. The authors found that the consistently estimated EOP are systematically affected by the two different combination approaches. The differences between the epoch and secular frames reach magnitudes of $23.6~\upmu \hbox {as}$ (0.73 mm) and $39.8~\upmu \hbox {as}$ (1.23 mm) for the x-pole and y-pole, respectively, in case of the combined solutions. For the SLR-only solutions, significant differences with amplitudes of $77.3~\upmu \hbox {as}$ (2.39 mm) can be found.  相似文献   

11.
A terrestrial survey, called the Geoid Slope Validation Survey of 2011 (GSVS11), encompassing leveling, GPS, astrogeodetic deflections of the vertical (DOV) and surface gravity was performed in the United States. The general purpose of that survey was to evaluate the current accuracy of gravimetric geoid models, and also to determine the impact of introducing new airborne gravity data from the ‘Gravity for the Redefinition of the American Vertical Datum’ (GRAV-D) project. More specifically, the GSVS11 survey was performed to determine whether or not the GRAV-D airborne gravimetry, flown at 11 km altitude, can reduce differential geoid error to below 1 cm in a low, flat gravimetrically uncomplicated region. GSVS11 comprises a 325 km traverse from Austin to Rockport in Southern Texas, and includes 218 GPS stations ( $\sigma _{\Delta h }= 0.4$ cm over any distance from 0.4 to 325 km) co-located with first-order spirit leveled orthometric heights ( $\sigma _{\Delta H }= 1.3$ cm end-to-end), including new surface gravimetry, and 216 astronomically determined vertical deflections $(\sigma _{\mathrm{DOV}}= 0.1^{\prime \prime })$ . The terrestrial survey data were compared in various ways to specific geoid models, including analysis of RMS residuals between all pairs of points on the line, direct comparison of DOVs to geoid slopes, and a harmonic analysis of the differences between the terrestrial data and various geoid models. These comparisons of the terrestrial survey data with specific geoid models showed conclusively that, in this type of region (low, flat) the geoid models computed using existing terrestrial gravity, combined with digital elevation models (DEMs) and GRACE and GOCE data, differential geoid accuracy of 1 to 3 cm (1 $\sigma )$ over distances from 0.4 to 325 km were currently being achieved. However, the addition of a contemporaneous airborne gravity data set, flown at 11 km altitude, brought the estimated differential geoid accuracy down to 1 cm over nearly all distances from 0.4 to 325 km.  相似文献   

12.
The integral formulas of the associated Legendre functions   总被引:1,自引:0,他引:1  
A new kind of integral formulas for ${\bar{P}_{n,m} (x)}$ is derived from the addition theorem about the Legendre Functions when n ? m is an even number. Based on the newly introduced integral formulas, the fully normalized associated Legendre functions can be directly computed without using any recursion methods that currently are often used in the computations. In addition, some arithmetic examples are computed with the increasing degree recursion and the integral methods introduced in the paper respectively, in order to compare the precisions and run-times of these two methods in computing the fully normalized associated Legendre functions. The results indicate that the precisions of the integral methods are almost consistent for variant x in computing ${\bar{P}_{n,m} (x)}$ , i.e., the precisions are independent of the choice of x on the interval [0,1]. In contrast, the precisions of the increasing degree recursion change with different values on the interval [0,1], particularly, when x tends to 1, the errors of computing ${\bar{P}_{n,m} (x)}$ by the increasing degree recursion become unacceptable when the degree becomes larger and larger. On the other hand, the integral methods cost more run-time than the increasing degree recursion. Hence, it is suggested that combinations of the integral method and the increasing degree recursion can be adopted, that is, the integral methods can be used as a replacement for the recursive initials when the recursion method become divergent.  相似文献   

13.
We show that the current levels of accuracy being achieved for the precise orbit determination (POD) of low-Earth orbiters demonstrate the need for the self-consistent treatment of tidal variations in the geocenter. Our study uses as an example the POD of the OSTM/Jason-2 satellite altimeter mission based upon Global Positioning System (GPS) tracking data. Current GPS-based POD solutions are demonstrating root-mean-square (RMS) radial orbit accuracy and precision of \({<}1\)  cm and 1 mm, respectively. Meanwhile, we show that the RMS of three-dimensional tidal geocenter variations is \({<}6\)  mm, but can be as large as 15 mm, with the largest component along the Earth’s spin axis. Our results demonstrate that GPS-based POD of Earth orbiters is best performed using GPS satellite orbit positions that are defined in a reference frame whose origin is at the center of mass of the entire Earth system, including the ocean tides. Errors in the GPS-based POD solutions for OSTM/Jason-2 of \({<}4\)  mm (3D RMS) and \({<}2\)  mm (radial RMS) are introduced when tidal geocenter variations are not treated consistently. Nevertheless, inconsistent treatment is measurable in the OSTM/Jason-2 POD solutions and manifests through degraded post-fit tracking data residuals, orbit precision, and relative orbit accuracy. For the latter metric, sea surface height crossover variance is higher by \(6~\hbox {mm}^{2}\) when tidal geocenter variations are treated inconsistently.  相似文献   

14.
Computation of broadcast ephemerides is a fundamental task in satellite navigation and positioning. The GPS constellation is composed of medium-earth-orbit (MEO) satellites, and therefore can employ a uniform parameter set to produce broadcast ephemerides. However, other navigation satellite systems such as Compass and IRNSS may include a mixture of inclined-geosynchronous-orbit (IGSO), geostationary-earth-orbit (GEO) and MEO satellites, requiring different parameter sets for each type of orbit. We analyze the variational characteristics of satellite ephemerides with respect to orbital elements; then present a method to design an optimal parameter set for broadcast ephemerides, and derive the parameter sets for IGSO, GEO, and MEO satellites. The computational complexities of the user algorithms for the optimal parameter sets are equivalent to that of the standard GPS user algorithm. Simulation and statistical analyses indicate that the optimal parameter set is $ \left\{ {\sqrt {A_{0} } ,e_{0} ,i_{0} ,\Upomega_{0} ,M_{0} ,\omega_{0} ,\dot{\Upomega },\dot{u},\dot{i},C_{\Upomega c3} ,C_{\Upomega s3} ,C_{uc2} ,C_{us2} ,C_{rc2} ,C_{rs2} } \right\} $ for IGSO and GEO satellites, and $ \left\{ {\sqrt {A_{0} } ,e_{0} ,i_{0} ,\Upomega_{0} ,M_{0} ,\omega_{0} ,\dot{\Upomega },\dot{u},\dot{i},C_{uc2} ,C_{us2} ,C_{rc2} ,C_{rs2} ,C_{ic2} ,C_{is2} } \right\} $ for MEO satellites.  相似文献   

15.
Deformations of radio telescopes used in geodetic and astrometric very long baseline interferometry (VLBI) observations belong to the class of systematic error sources which require correction in data analysis. In this paper we present a model for all path length variations in the geometrical optics of radio telescopes which are due to gravitational deformation. The Effelsberg 100 m radio telescope of the Max Planck Institute for Radio Astronomy, Bonn, Germany, has been surveyed by various terrestrial methods. Thus, all necessary information that is needed to model the path length variations is available. Additionally, a ray tracing program has been developed which uses as input the parameters of the measured deformations to produce an independent check of the theoretical model. In this program as well as in the theoretical model, the illumination function plays an important role because it serves as the weighting function for the individual path lengths depending on the distance from the optical axis. For the Effelsberg telescope, the biggest contribution to the total path length variations is the bending of the main beam located along the elevation axis which partly carries the weight of the paraboloid at its vertex. The difference in total path length is almost \(-\) 100 mm when comparing observations at 90 \(^\circ \) and at 0 \(^\circ \) elevation angle. The impact of the path length corrections is validated in a global VLBI analysis. The application of the correction model leads to a change in the vertical position of \(+120\)  mm. This is more than the maximum path length, but the effect can be explained by the shape of the correction function.  相似文献   

16.
Reducing the draconitic errors in GNSS geodetic products   总被引:2,自引:2,他引:0  
Systematic errors at harmonics of the GPS draconitic year have been found in diverse GPS-derived geodetic products like the geocenter $Z$ -component, station coordinates, $Y$ -pole rate and orbits (i.e. orbit overlaps). The GPS draconitic year is the repeat period of the GPS constellation w.r.t. the Sun which is about 351 days. Different error sources have been proposed which could generate these spurious signals at the draconitic harmonics. In this study, we focus on one of these error sources, namely the radiation pressure orbit modeling deficiencies. For this purpose, three GPS+GLONASS solutions of 8 years (2004–2011) were computed which differ only in the solar radiation pressure (SRP) and satellite attitude models. The models employed in the solutions are: (1) the CODE (5-parameter) radiation pressure model widely used within the International GNSS Service community, (2) the adjustable box-wing model for SRP impacting GPS (and GLONASS) satellites, and (3) the adjustable box-wing model upgraded to use non-nominal yaw attitude, specially for satellites in eclipse seasons. When comparing the first solution with the third one we achieved the following in the GNSS geodetic products. Orbits: the draconitic errors in the orbit overlaps are reduced for the GPS satellites in all the harmonics on average 46, 38 and 57 % for the radial, along-track and cross-track components, while for GLONASS satellites they are mainly reduced in the cross-track component by 39 %. Geocenter $Z$ -component: all the odd draconitic harmonics found when the CODE model is used show a very important reduction (almost disappearing with a 92 % average reduction) with the new radiation pressure models. Earth orientation parameters: the draconitic errors are reduced for the $X$ -pole rate and especially for the $Y$ -pole rate by 24 and 50 % respectively. Station coordinates: all the draconitic harmonics (except the 2nd harmonic in the North component) are reduced in the North, East and Height components, with average reductions of 41, 39 and 35 % respectively. This shows, that part of the draconitic errors currently found in GNSS geodetic products are definitely induced by the CODE radiation pressure orbit modeling deficiencies.  相似文献   

17.
Earth orientation parameters estimated from VLBI during the CONT11 campaign   总被引:1,自引:1,他引:0  
In this paper we investigate the accuracy of the earth orientation parameters (EOP) estimated from the continuous VLBI campaign CONT11. We first estimated EOP with daily resolution and compared these to EOP estimated from GNSS data. We find that the WRMS differences are about 31  $\upmu $ as for polar motion and 7  $\upmu $ s for length of day. This is about the precision we could expect, based on Monte Carlo simulations and the results of the previous CONT campaigns. We also estimated EOP with hourly resolution to study the sub-diurnal variations. The results confirm the results of previous studies, showing that the current IERS model for high-frequency EOP variations does not explain all the sub-diurnal variations seen in the estimated time series. We then compared our results to various empirical high-frequency EOP models. However, we did not find that any of these gave any unambiguous improvement. Several simulations testing the impact of various aspects of, e.g. the observing network were also made. For example, we made simulations assuming that all CONT11 stations were equipped with fast VLBI2010 antennas. We found that the WRMS error decreased by about a factor five compared to the current VLBI system. Furthermore, the simulations showed that it is very important to have a homogenous global distribution of the stations for achieving the highest precision for the EOP.  相似文献   

18.
We report on the susceptibility of the Scintrex CG-5 relative gravimeters to tilting, that is the tendency of the instrument of providing incorrect readings after being tilted (even by small angles) for a moderate period of time. Tilting of the instrument can occur when in transit between sites usually on the backseat of a car even using the specially designed transport case. Based on a series of experiments with different instruments, we demonstrate that the readings may be offset by tens of $\upmu $ Gal. In addition, it may take hours before the first reliable readings can be taken, with the actual time depending on how long the instrument had been tilted. This sensitivity to tilt in combination with the long time required for the instrument to provide reliable readings has not yet been reported in the literature and is not addressed adequately in the Scintrex CG-5 user manual. In particular, the inadequate instrument state cannot easily be detected by checking the readings during the observation or by reviewing the final data before leaving a site, precautions suggested by Scintrex Ltd. In regional surveys with car transportation over periods of tens of minutes to hours, the gravity measurements can be degraded by some 10  $\upmu $ Gal. To obtain high-quality results in line with the CG-5 specifications, the gravimeters must remain in upright position to within a few degrees during transits. This requirement may often be unrealistic during field observations, particularly when observing in hilly terrain or when walking with the instrument in a backpack.  相似文献   

19.
The present paper deals with the least-squares adjustment where the design matrix (A) is rank-deficient. The adjusted parameters \(\hat x\) as well as their variance-covariance matrix ( \(\sum _{\hat x} \) ) can be obtained as in the “standard” adjustment whereA has the full column rank, supplemented with constraints, \(C\hat x = w\) , whereC is the constraint matrix andw is sometimes called the “constant vector”. In this analysis only the inner adjustment constraints are considered, whereC has the full row rank equal to the rank deficiency ofA, andAC T =0. Perhaps the most important outcome points to the three kinds of results
  1. A general least-squares solution where both \(\hat x\) and \(\sum _{\hat x} \) are indeterminate corresponds tow=arbitrary random vector.
  2. The minimum trace (least-squares) solution where \(\hat x\) is indeterminate but \(\sum _{\hat x} \) is detemined (and trace \(\sum _{\hat x} \) corresponds tow=arbitrary constant vector.
  3. The minimum norm (least-squares) solution where both \(\hat x\) and \(\sum _{\hat x} \) are determined (and norm \(\hat x\) , trace \(\sum _{\hat x} \) corresponds tow?0
  相似文献   

20.
Estimation of variance and covariance components   总被引:3,自引:2,他引:3  
  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号