Robust Uncertainty Quantification and Scalable Computation for Computer Models with Massive Output

نویسندگان

  • Mengyang Gu
  • Robert L. Wolpert
  • Surya Tokdar
  • Barbara Engelhardt
چکیده

Robust Uncertainty Quantification and Scalable Computation for Computer Models with Massive Output by Mengyang Gu Department of Statistical Science Duke University Date: Approved: James O. Berger, Supervisor Robert L. Wolpert Surya Tokdar Barbara Engelhardt An abstract of a dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Department of Statistical Science in the Graduate School of Duke University 2016 Copyright c © 2016 by Mengyang Gu All rights reserved except the rights granted by the Creative Commons Attribution-Noncommercial Licence Abstract Computer models have been widely used to reproduce the behavior of engineering, physical, biological and human processes. The rapid development of technology has empowered scientists and engineers to implement large-scale simulations via computer models and to collect real data from different kinds of resources, with the ultimate goal of predicting real processes through modeling. To achieve this typically requires a host of interactions with data and statistics, a process that has come to be called uncertainty quantification (UQ). This thesis develops statistical models focusing on two aspects of the problem in UQ: computational feasibility for huge functional data and robust parameter estimation in fitting models. To achieve these two goals, new techniques, theories and numerical procedures are developed and studied. Chapter 1 frames the issue of modeling data from computer experiments and introduces the Gaussian stochastic process (GaSP) emulator as a crucial step in UQ. Chapter 2 provides a practical approach for simultaneously emulating/approximating computer models that produce massive output; for instance, output over a huge range of space-time coordinates (as necessary for the discussed application of hazard quantification for the Soufrière Hills volcano in Montserrat island). Chapter 3 and Chapter 4 are both about the parameter estimation problem for the GaSP emulator. Chapter 3 provides new criteria for parameter estimation, called ‘robustness parameter estimation criteria’. Properties of the reference prior are studied in a general ivComputer models have been widely used to reproduce the behavior of engineering, physical, biological and human processes. The rapid development of technology has empowered scientists and engineers to implement large-scale simulations via computer models and to collect real data from different kinds of resources, with the ultimate goal of predicting real processes through modeling. To achieve this typically requires a host of interactions with data and statistics, a process that has come to be called uncertainty quantification (UQ). This thesis develops statistical models focusing on two aspects of the problem in UQ: computational feasibility for huge functional data and robust parameter estimation in fitting models. To achieve these two goals, new techniques, theories and numerical procedures are developed and studied. Chapter 1 frames the issue of modeling data from computer experiments and introduces the Gaussian stochastic process (GaSP) emulator as a crucial step in UQ. Chapter 2 provides a practical approach for simultaneously emulating/approximating computer models that produce massive output; for instance, output over a huge range of space-time coordinates (as necessary for the discussed application of hazard quantification for the Soufrière Hills volcano in Montserrat island). Chapter 3 and Chapter 4 are both about the parameter estimation problem for the GaSP emulator. Chapter 3 provides new criteria for parameter estimation, called ‘robustness parameter estimation criteria’. Properties of the reference prior are studied in a general iv setting and a new robust estimation procedure is proposed based on the estimation from marginal posterior modes with optimal parameterizations. Chapter 4 introduces a new class of priors – called jointly robust priors – for the GaSP model where inert inputs (inputs that barely affect the computer model output) are present. This prior has many of the good properties of the reference prior and is more computationally tractable. Chapter 5 discusses another problem with big functional data, in which the number of observations in a function is large. An exact algorithm that can compute the likelihood of the GaSP model linearly in time is introduced, and regression, separable GaSP models and nonseparable GaSP models are discussed in a unified framework. Finally, Chapter 6 provides some concluding remarks and describes possible future work.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Measuring robust overall profit efficiency with uncertainty in input and output price vectors

The classic overall profit needs precise information of inputs, outputs, inputs and outputs price vectors. In real word, all data are not certain. Therefore, in this case, stochastic and fuzzy methods use for measuring overall profit efficiency. These methods require more information about the data such as probability distribution function or data membership function, which in some cases may no...

متن کامل

Scalable Approach to Uncertainty Quantification and Robust Design of Interconnected Dynamical Systems

Development of robust dynamical systems and networks such as autonomous aircraft systems capable of accomplishing complex missions faces challenges due to the dynamically evolving uncertainties coming from model uncertainties, necessity to operate in a hostile cluttered urban environment, and the distributed and dynamic nature of the communication and computation resources. Model-based robust d...

متن کامل

Relative Efficiency Measurement of Banks Using Network DEA Model in Uncertainty Situation

Traditional DEA method considered decision making units (DMUs) as a black box, regardless of their internal structure and appraisal performance with respect to the final inputs and outputs of the units. However, in many real systems we have internal structure. For this reason, network DEA models have been developed. Parallel network DEA models are a special variation which inputs of unit alloca...

متن کامل

Dynamic configuration and collaborative scheduling in supply chains based on scalable multi-agent architecture

Due to diversified and frequently changing demands from customers, technological advances and global competition, manufacturers rely on collaboration with their business partners to share costs, risks and expertise. How to take advantage of advancement of technologies to effectively support operations and create competitive advantage is critical for manufacturers to survive. To respond to these...

متن کامل

DPML-Risk: An Efficient Algorithm for Image Registration

Targets and objects registration and tracking in a sequence of images play an important role in various areas. One of the methods in image registration is feature-based algorithm which is accomplished in two steps. The first step includes finding features of sensed and reference images. In this step, a scale space is used to reduce the sensitivity of detected features to the scale changes. Afterw...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016