Expanded uncertainty quantification in inverse problems: Hierarchical Bayes and empirical Bayes

Geophysics ◽  
2004 ◽  
Vol 69 (4) ◽  
pp. 1005-1016 ◽  
Author(s):  
Alberto Malinverno ◽  
Victoria A. Briggs

A common way to account for uncertainty in inverse problems is to apply Bayes' rule and obtain a posterior distribution of the quantities of interest given a set of measurements. A conventional Bayesian treatment, however, requires assuming specific values for parameters of the prior distribution and of the distribution of the measurement errors (e.g., the standard deviation of the errors). In practice, these parameters are often poorly known a priori, and choosing a particular value is often problematic. Moreover, the posterior uncertainty is computed assuming that these parameters are fixed; if they are not well known a priori, the posterior uncertainties have dubious value. This paper describes extensions to the conventional Bayesian treatment that assign uncertainty to the parameters defining the prior distribution and the distribution of the measurement errors. These extensions are known in the statistical literature as “empirical Bayes” and “hierarchical Bayes.” We demonstrate the practical application of these approaches to a simple linear inverse problem: using seismic traveltimes measured by a receiver in a well to infer compressional wave slowness in a 1D earth model. These procedures do not require choosing fixed values for poorly known parameters and, at most, need a realistic range (e.g., a minimum and maximum value for the standard deviation of the measurement errors). Inversion is thus made easier for general users, who are not required to set parameters they know little about.

2002 ◽  
Vol 59 (5) ◽  
pp. 854-864 ◽  
Author(s):  
Daniel G Kehler ◽  
Ransom A Myers ◽  
Christopher A Field

We investigated the consequences of measurement errors in spawner abundance on the estimation of alpha(tilde), the maximum lifetime reproductive rate, in the Ricker spawner-recruit model. This rate is equivalent to the slope at the origin for the Ricker model, which is crucial for the estimation of management reference points. Simulation results indicate that measurement error in spawner abundance does not lead, in general, to an overestimate of alpha(tilde), as had been previously thought. Instead, the direction, and to some extent the magnitude, of the bias is related to the range of spawner values, and we provide an analytical explanation for this phenomenon. For stocks that have been heavily exploited and have low spawner values, measurement error tends to result in an overestimate of log alpha(tilde), usually by [Formula: see text]5%. Less exploited stocks with large spawner values will have their log alpha(tilde) underestimated, commonly by –15 to –25%. Results from data-based simulations indicate that under certain conditions, empirical Bayes estimates for the distributions for alpha(tilde) among stocks, i.e., the prior distribution, can still be relatively unbiased in the presence of measurement error.


2018 ◽  
Vol 16 (2) ◽  
pp. 142-153 ◽  
Author(s):  
Kristen M Cunanan ◽  
Alexia Iasonos ◽  
Ronglai Shen ◽  
Mithat Gönen

Background: In the era of targeted therapies, clinical trials in oncology are rapidly evolving, wherein patients from multiple diseases are now enrolled and treated according to their genomic mutation(s). In such trials, known as basket trials, the different disease cohorts form the different baskets for inference. Several approaches have been proposed in the literature to efficiently use information from all baskets while simultaneously screening to find individual baskets where the drug works. Most proposed methods are developed in a Bayesian paradigm that requires specifying a prior distribution for a variance parameter, which controls the degree to which information is shared across baskets. Methods: A common approach used to capture the correlated binary endpoints across baskets is Bayesian hierarchical modeling. We evaluate a Bayesian adaptive design in the context of a non-randomized basket trial and investigate three popular prior specifications: an inverse-gamma prior on the basket-level variance, a uniform prior and half-t prior on the basket-level standard deviation. Results: From our simulation study, we can see that the inverse-gamma prior is highly sensitive to the input hyperparameters. When the prior mean value of the variance parameter is set to be near zero [Formula: see text], this can lead to unacceptably high false-positive rates [Formula: see text] in some scenarios. Thus, use of this prior requires a fully comprehensive sensitivity analysis before implementation. Alternatively, we see that a prior that places sufficient mass in the tail, such as the uniform or half-t prior, displays desirable and robust operating characteristics over a wide range of prior specifications, with the caveat that the upper bound of the uniform prior and the scale parameter of the half-t prior must be larger than 1. Conclusion: Based on the simulation results, we recommend that those involved in designing basket trials that implement hierarchical modeling avoid using a prior distribution that places a majority of the density mass near zero for the variance parameter. Priors with this property force the model to share information regardless of the true efficacy configuration of the baskets. Many commonly used inverse-gamma prior specifications have this undesirable property. We recommend to instead consider the more robust uniform prior or half-t prior on the standard deviation.


2015 ◽  
Vol 8 (2) ◽  
pp. 941-963 ◽  
Author(s):  
T. Vlemmix ◽  
F. Hendrick ◽  
G. Pinardi ◽  
I. De Smedt ◽  
C. Fayt ◽  
...  

Abstract. A 4-year data set of MAX-DOAS observations in the Beijing area (2008–2012) is analysed with a focus on NO2, HCHO and aerosols. Two very different retrieval methods are applied. Method A describes the tropospheric profile with 13 layers and makes use of the optimal estimation method. Method B uses 2–4 parameters to describe the tropospheric profile and an inversion based on a least-squares fit. For each constituent (NO2, HCHO and aerosols) the retrieval outcomes are compared in terms of tropospheric column densities, surface concentrations and "characteristic profile heights" (i.e. the height below which 75% of the vertically integrated tropospheric column density resides). We find best agreement between the two methods for tropospheric NO2 column densities, with a standard deviation of relative differences below 10%, a correlation of 0.99 and a linear regression with a slope of 1.03. For tropospheric HCHO column densities we find a similar slope, but also a systematic bias of almost 10% which is likely related to differences in profile height. Aerosol optical depths (AODs) retrieved with method B are 20% high compared to method A. They are more in agreement with AERONET measurements, which are on average only 5% lower, however with considerable relative differences (standard deviation ~ 25%). With respect to near-surface volume mixing ratios and aerosol extinction we find considerably larger relative differences: 10 ± 30, −23 ± 28 and −8 ± 33% for aerosols, HCHO and NO2 respectively. The frequency distributions of these near-surface concentrations show however a quite good agreement, and this indicates that near-surface concentrations derived from MAX-DOAS are certainly useful in a climatological sense. A major difference between the two methods is the dynamic range of retrieved characteristic profile heights which is larger for method B than for method A. This effect is most pronounced for HCHO, where retrieved profile shapes with method A are very close to the a priori, and moderate for NO2 and aerosol extinction which on average show quite good agreement for characteristic profile heights below 1.5 km. One of the main advantages of method A is the stability, even under suboptimal conditions (e.g. in the presence of clouds). Method B is generally more unstable and this explains probably a substantial part of the quite large relative differences between the two methods. However, despite a relatively low precision for individual profile retrievals it appears as if seasonally averaged profile heights retrieved with method B are less biased towards a priori assumptions than those retrieved with method A. This gives confidence in the result obtained with method B, namely that aerosol extinction profiles tend on average to be higher than NO2 profiles in spring and summer, whereas they seem on average to be of the same height in winter, a result which is especially relevant in relation to the validation of satellite retrievals.


2020 ◽  
Vol 28 (3) ◽  
pp. 449-463 ◽  
Author(s):  
Natalia P. Bondarenko ◽  
Chung-Tsun Shieh

AbstractIn this paper, partial inverse problems for the quadratic pencil of Sturm–Liouville operators on a graph with a loop are studied. These problems consist in recovering the pencil coefficients on one edge of the graph (a boundary edge or the loop) from spectral characteristics, while the coefficients on the other edges are known a priori. We obtain uniqueness theorems and constructive solutions for partial inverse problems.


2019 ◽  
Vol 27 (3) ◽  
pp. 317-340 ◽  
Author(s):  
Max Kontak ◽  
Volker Michel

Abstract In this work, we present the so-called Regularized Weak Functional Matching Pursuit (RWFMP) algorithm, which is a weak greedy algorithm for linear ill-posed inverse problems. In comparison to the Regularized Functional Matching Pursuit (RFMP), on which it is based, the RWFMP possesses an improved theoretical analysis including the guaranteed existence of the iterates, the convergence of the algorithm for inverse problems in infinite-dimensional Hilbert spaces, and a convergence rate, which is also valid for the particular case of the RFMP. Another improvement is the cancellation of the previously required and difficult to verify semi-frame condition. Furthermore, we provide an a-priori parameter choice rule for the RWFMP, which yields a convergent regularization. Finally, we will give a numerical example, which shows that the “weak” approach is also beneficial from the computational point of view. By applying an improved search strategy in the algorithm, which is motivated by the weak approach, we can save up to 90  of computation time in comparison to the RFMP, whereas the accuracy of the solution does not change as much.


Author(s):  
Natalia Bondarenko ◽  
Chung-Tsun Shieh

In this paper, inverse spectral problems for Sturm–Liouville operators on a tree (a graph without cycles) are studied. We show that if the potential on an edge is known a priori, then b – 1 spectral sets uniquely determine the potential functions on a tree with b external edges. Constructive solutions, based on the method of spectral mappings, are provided for the considered inverse problems.


2014 ◽  
Vol 14 (23) ◽  
pp. 12897-12914 ◽  
Author(s):  
J. S. Wang ◽  
S. R. Kawa ◽  
J. Eluszkiewicz ◽  
D. F. Baker ◽  
M. Mountain ◽  
...  

Abstract. Top–down estimates of the spatiotemporal variations in emissions and uptake of CO2 will benefit from the increasing measurement density brought by recent and future additions to the suite of in situ and remote CO2 measurement platforms. In particular, the planned NASA Active Sensing of CO2 Emissions over Nights, Days, and Seasons (ASCENDS) satellite mission will provide greater coverage in cloudy regions, at high latitudes, and at night than passive satellite systems, as well as high precision and accuracy. In a novel approach to quantifying the ability of satellite column measurements to constrain CO2 fluxes, we use a portable library of footprints (surface influence functions) generated by the Stochastic Time-Inverted Lagrangian Transport (STILT) model in combination with the Weather Research and Forecasting (WRF) model in a regional Bayesian synthesis inversion. The regional Lagrangian particle dispersion model framework is well suited to make use of ASCENDS observations to constrain weekly fluxes in North America at a high resolution, in this case at 1° latitude × 1° longitude. We consider random measurement errors only, modeled as a function of the mission and instrument design specifications along with realistic atmospheric and surface conditions. We find that the ASCENDS observations could potentially reduce flux uncertainties substantially at biome and finer scales. At the grid scale and weekly resolution, the largest uncertainty reductions, on the order of 50%, occur where and when there is good coverage by observations with low measurement errors and the a priori uncertainties are large. Uncertainty reductions are smaller for a 1.57 μm candidate wavelength than for a 2.05 μm wavelength, and are smaller for the higher of the two measurement error levels that we consider (1.0 ppm vs. 0.5 ppm clear-sky error at Railroad Valley, Nevada). Uncertainty reductions at the annual biome scale range from ~40% to ~75% across our four instrument design cases and from ~65% to ~85% for the continent as a whole. Tests suggest that the quantitative results are moderately sensitive to assumptions regarding a priori uncertainties and boundary conditions. The a posteriori flux uncertainties we obtain, ranging from 0.01 to 0.06 Pg C yr−1 across the biomes, would meet requirements for improved understanding of long-term carbon sinks suggested by a previous study.


2021 ◽  
Author(s):  
Natacha Galmiche ◽  
Nello Blaser ◽  
Morten Brun ◽  
Helwig Hauser ◽  
Thomas Spengler ◽  
...  

<p>Probability distributions based on ensemble forecasts are commonly used to assess uncertainty in weather prediction. However, interpreting these distributions is not trivial, especially in the case of multimodality with distinct likely outcomes. The conventional summary employs mean and standard deviation across ensemble members, which works well for unimodal, Gaussian-like distributions. In the case of multimodality this misleads, discarding crucial information. </p><p>We aim at combining previously developed clustering algorithms in machine learning and topological data analysis to extract useful information such as the number of clusters in an ensemble. Given the chaotic behaviour of the atmosphere, machine learning techniques can provide relevant results even if no, or very little, a priori information about the data is available. In addition, topological methods that analyse the shape of the data can make results explainable.</p><p>Given an ensemble of univariate time series, a graph is generated whose edges and vertices represent clusters of members, including additional information for each cluster such as the members belonging to them, their uncertainty, and their relevance according to the graph. In the case of multimodality, this approach provides relevant and quantitative information beyond the commonly used mean and standard deviation approach that helps to further characterise the predictability.</p>


Geophysics ◽  
2021 ◽  
pp. 1-44
Author(s):  
Ujjal K. Borah ◽  
Prasanta K. Patro

Large man-made water-reservoirs promote fluid diffusion and cause critically stressed fault zones underneath to trigger earthquakes. Electrical resistivity is a crucial property to investigate such fluid-filled fault zones. We, therefore, carry out magnetotelluric (MT) investigation to explore an intra-plate earthquake zone, which is related to artificial reservoir triggered seismicity. However, due to surface access restrictions, our dataset has a gap in coverage in the middle part of the study area. This data gap region coincides with the earthquake hypocenter distribution in that intra-plate earthquake zone. Therefore, it is vital to fill the data gap to get the electrical signature of the active seismic zone. To compensate for the data gap, we develop a relation that connects resistivity with the ratio of seismic P- to S-wave velocity ( VP/ VS). Utilizing this relation, we estimate a priori resistivity distribution in the data gap region from known vp/vs values during inversion to compensate for the data gap. A comparison study of the root mean square (RMS) misfits of inversion outputs (with and without data gap filled) proves the effectiveness of the established relation. The inversion outputs obtained using the established relation brings out fault signatures in the data gap region. To examine the reliability and accuracy of these fault signatures, we occupy a portion of the data gap with new MT sites. We compare the inversion output from this new setup with the inversion output obtained from the established relation and observe that the electrical signatures in both outputs are spatially correlated. Further, a synthetic test on a similar earth model establishes the credibility and robustness of the derived relation.


1976 ◽  
Vol 66 (1) ◽  
pp. 173-187
Author(s):  
Ray Buland

abstract A complete reexamination of Geiger's method in the light of modern numerical analysis indicates that numerical stability can be insured by use of the QR algorithm and the convergence domain considerably enlarged by the introduction of step-length damping. In order to make the maximum use of all data, the method is developed assuming a priori estimates of the statistics of the random errors at each station. Numerical experiments indicate that the bulk of the joint probability density of the location parameters is in the linear region allowing simple estimates of the standard errors of the parameters. The location parameters are found to be distributed as one minus chi squared with m degrees of freedom, where m is the number of parameters, allowing the simple construction of confidence levels. The use of the chi-squared test with n-m degrees of freedom, where n is the number of data, is introduced as a means of qualitatively evaluating the correctness of the earth model.


Sign in / Sign up

Export Citation Format

Share Document