Open Access. Powered by Scholars. Published by Universities.®

Statistical Methodology Commons

Open Access. Powered by Scholars. Published by Universities.®

1,113 Full-Text Articles 1,506 Authors 603,045 Downloads 125 Institutions

All Articles in Statistical Methodology

Faceted Search

1,113 full-text articles. Page 30 of 38.

Instrumental Variable Analyses: Exploiting Natural Randomness To Understand Causal Mechanisms, Theodore Iwashyna, Edward Kennedy 2012 University of Michigan

Instrumental Variable Analyses: Exploiting Natural Randomness To Understand Causal Mechanisms, Theodore Iwashyna, Edward Kennedy

Edward H. Kennedy

Instrumental variable analysis is a technique commonly used in the social sciences to provide evidence that a treatment causes an outcome, as contrasted with evidence that a treatment is merely associated with differences in an outcome. To extract such strong evidence from observational data, instrumental variable analysis exploits situations where some degree of randomness affects how patients are selected for a treatment. An instrumental variable is a characteristic of the world that leads some people to be more likely to get the specific treatment we want to study but does not otherwise change thosepatients’ outcomes. This seminar explains, in nonmathematical …


Theory And Methods For Gini Coefficients Partitioned By Quantile Range, Chaitra Nagaraja 2012 Fordham University

Theory And Methods For Gini Coefficients Partitioned By Quantile Range, Chaitra Nagaraja

Chaitra H Nagaraja

The Gini coefficient is frequently used to measure inequality in populations. However, it is possible that inequality levels may change over time differently for disparate subgroups which cannot be detected with population-level estimates only. Therefore, it may be informative to examine inequality separately for these segments. The case where the population is split into two segments based on non-overlapping quantile ranges is examined. Asymptotic theory is derived and practical methods to estimate standard errors and construct confidence intervals using resampling methods are developed. An application to per capita income across census tracts using American Community Survey data is considered.


A Comparison Of Periodic Autoregressive And Dynamic Factor Models In Intraday Energy Demand Forecasting, Thomas Mestekemper, Goeran Kauermann, Michael Smith 2012 Melbourne Business School

A Comparison Of Periodic Autoregressive And Dynamic Factor Models In Intraday Energy Demand Forecasting, Thomas Mestekemper, Goeran Kauermann, Michael Smith

Michael Stanley Smith

We suggest a new approach for forecasting energy demand at an intraday resolution. Demand in each intraday period is modeled using semiparametric regression smoothing to account for calendar and weather components. Residual serial dependence is captured by one of two multivariate stationary time series models, with dimension equal to the number of intraday periods. These are a periodic autoregression and a dynamic factor model. We show the benefits of our approach in the forecasting of district heating demand in a steam network in Germany and aggregate electricity demand in the state of Victoria, Australia. In both studies, accounting for weather …


An Overview Of Targeted Maximum Likelihood Estimation, Susan Gruber 2012 Harvard School of Public Health

An Overview Of Targeted Maximum Likelihood Estimation, Susan Gruber

Susan Gruber

These slides provide an introduction to targeted maximum likelihood estimation in a point treatment setting.


Bayesian Approaches To Copula Modelling, Michael S. Smith 2012 Melbourne Business School

Bayesian Approaches To Copula Modelling, Michael S. Smith

Michael Stanley Smith

Copula models have become one of the most widely used tools in the applied modelling of multivariate data. Similarly, Bayesian methods are increasingly used to obtain efficient likelihood-based inference. However, to date, there has been only limited use of Bayesian approaches in the formulation and estimation of copula models. This article aims to address this shortcoming in two ways. First, to introduce copula models and aspects of copula theory that are especially relevant for a Bayesian analysis. Second, to outline Bayesian approaches to formulating and estimating copula models, and their advantages over alternative methods. Copulas covered include Archimedean, copulas constructed …


Bayesian Methods For Expression-Based Integration, Elizabeth M. Jennings, Jeffrey S. Morris, Raymond J. Carroll, Ganiraju C. Manyam, Veera Baladandayuthapani 2012 Texas A&M University

Bayesian Methods For Expression-Based Integration, Elizabeth M. Jennings, Jeffrey S. Morris, Raymond J. Carroll, Ganiraju C. Manyam, Veera Baladandayuthapani

Jeffrey S. Morris

We propose methods to integrate data across several genomic platforms using a hierarchical Bayesian analysis framework that incorporates the biological relationships among the platforms to identify genes whose expression is related to clinical outcomes in cancer. This integrated approach combines information across all platforms, leading to increased statistical power in finding these predictive genes, and further provides mechanistic information about the manner in which the gene affects the outcome. We demonstrate the advantages of the shrinkage estimation used by this approach through a simulation, and finally, we apply our method to a Glioblastoma Multiforme dataset and identify several genes potentially …


Quantifying Temporal Correlations: A Test-Retest Evaluation Of Functional Connectivity In Resting-State Fmri, Mark Fiecas, Hernando Ombao, Dan van Lunen, Richard Baumgartner, Alexandre Coimbra, Dai Feng 2012 University of California - San Diego

Quantifying Temporal Correlations: A Test-Retest Evaluation Of Functional Connectivity In Resting-State Fmri, Mark Fiecas, Hernando Ombao, Dan Van Lunen, Richard Baumgartner, Alexandre Coimbra, Dai Feng

Mark Fiecas

There have been many interpretations of functional connectivity and proposed measures of temporal correlations between BOLD signals across different brain areas. These interpretations yield from many studies on functional connectivity using resting-state fMRI data that have emerged in recent years. However, not all of these studies used the same metrics for quantifying the temporal correlations between brain regions. In this paper, we use a public-domain test–retest resting-state fMRI data set to perform a systematic investigation of the stability of the metrics that are often used in resting-state functional connectivity (FC) studies. The fMRI data set was collected across three different …


Methods For Evaluating Prediction Performance Of Biomarkers And Tests, Margaret S. Pepe PhD, Holly Janes PhD 2012 Fred Hutchinson Cancer Research Center

Methods For Evaluating Prediction Performance Of Biomarkers And Tests, Margaret S. Pepe Phd, Holly Janes Phd

Margaret S Pepe PhD

This chapter describes and critiques methods for evaluating the performance of markers to predict risk of a current or future clinical outcome. We consider three criteria that are important for evaluating a risk model: calibration, benefit for decision making and accurate classification. We also describe and discuss a variety of summary measures in common use for quantifying predictive information such as the area under the ROC curve and R-squared. The roles and problems with recently proposed risk reclassification approaches are discussed in detail.


Group Testing Regression Models, Boan Zhang 2012 University of Nebraska-Lincoln

Group Testing Regression Models, Boan Zhang

Department of Statistics: Dissertations, Theses, and Student Work

Group testing, where groups of individual specimens are composited to test for the presence or absence of a disease (or some other binary characteristic), is a procedure commonly used to reduce the costs of screening a large number of individuals. Statistical research in group testing has traditionally focused on a homogeneous population, where individuals are assumed to have the same probability of having a disease. However, individuals often have different risks of positivity, so recent research has examined regression models that allow for heterogeneity among individuals within the population. This dissertation focuses on two problems involving group testing regression models. …


Obtaining Critical Values For Test Of Markov Regime Switching, Douglas G. Steigerwald, Valerie Bostwick 2012 University of California, Santa Barbara

Obtaining Critical Values For Test Of Markov Regime Switching, Douglas G. Steigerwald, Valerie Bostwick

Douglas G. Steigerwald

For Markov regime-switching models, testing for the possible presence of more than one regime requires the use of a non-standard test statistic. Carter and Steigerwald (forthcoming, Journal of Econometric Methods) derive in detail the analytic steps needed to implement the test ofMarkov regime-switching proposed by Cho and White (2007, Econometrica). We summarize the implementation steps and address the computational issues that arise. A new command to compute regime-switching critical values, rscv, is introduced and presented in the context of empirical research.


Quest For Continuous Improvement: Gathering Feedback And Data Through Multiple Methods To Evaluate And Improve A Library’S Discovery Tool, Jeanne M. Brown 2012 University of Nevada, Las Vegas

Quest For Continuous Improvement: Gathering Feedback And Data Through Multiple Methods To Evaluate And Improve A Library’S Discovery Tool, Jeanne M. Brown

Library Faculty Presentations

Summon at UNLV

  • Implemented fall 2011: a web-scale discovery tool
  • Expectations for Summon
  • Continuous Summon Improvement (CSI)Group

The environment

  • User changes
  • Library changes
  • Vendor changes
  • Product changes
  • Complex information environment
  • Change + complexity = need to assess using multiple streams of feedback


A Doubling Technique For The Power Method Transformations, Mohan D. Pant, Todd C. Headrick 2012 University of Texas at Arlington

A Doubling Technique For The Power Method Transformations, Mohan D. Pant, Todd C. Headrick

Mohan Dev Pant

Power method polynomials are used for simulating non-normal distributions with specified product moments or L-moments. The power method is capable of producing distributions with extreme values of skew (L-skew) and kurtosis (L-kurtosis). However, these distributions can be extremely peaked and thus not representative of real-world data. To obviate this problem, two families of distributions are introduced based on a doubling technique with symmetric standard normal and logistic power method distributions. The primary focus of the methodology is in the context of L-moment theory. As such, L-moment based systems of equations are derived for simulating univariate and multivariate non-normal distributions with …


Finding A Better Confidence Interval For A Single Regression Changepoint Using Different Bootstrap Confidence Interval Procedures, Bodhipaksha Thilakarathne 2012 Georgia Southern University

Finding A Better Confidence Interval For A Single Regression Changepoint Using Different Bootstrap Confidence Interval Procedures, Bodhipaksha Thilakarathne

Electronic Theses and Dissertations

Recently a number of papers have been published in the area of regression changepoints but there is not much literature concerning confidence intervals for regression changepoints. The purpose of this paper is to find a better bootstrap confidence interval for a single regression changepoint. ("Better" confidence interval means having a minimum length and coverage probability which is close to a chosen significance level). Several methods will be used to find bootstrap confidence intervals. Among those methods a better confidence interval will be presented.


Adventures In Library Salary Surveys, Scott L. Schaffer 2012 University of Vermont

Adventures In Library Salary Surveys, Scott L. Schaffer

UVM Libraries Conference Day

Salary surveys are an important tool for the library community and the administrators and boards responsible for the oversight of libraries. However, such assessments must be constructed and analyzed with great care. The Vermont Library Association Personnel Committee has conducted three salary surveys over the past several years, one focusing on academic libraries and two on public libraries. Significant issues have included confidentiality, participation rate, definitions, length and difficulty of questions, collection of data, and representativeness. Suggestions and lessons learned will be shared.


An L-Moment-Based Analog For The Schmeiser-Deutsch Class Of Distributions, Todd C. Headrick, Mohan D. Pant 2012 Southern Illinois University Carbondale

An L-Moment-Based Analog For The Schmeiser-Deutsch Class Of Distributions, Todd C. Headrick, Mohan D. Pant

Mohan Dev Pant

This paper characterizes the conventional moment-based Schmeiser-Deutsch (S-D) class of distributions through the method of L-moments. The system can be used in a variety of settings such as simulation or modeling various processes. A procedure is also described for simulating S-D distributions with specified L-moments and L-correlations. The Monte Carlo results presented in this study indicate that the estimates of L-skew, L-kurtosis, and L-correlation associated with the S-D class of distributions are substantially superior to their corresponding conventional product-moment estimators in terms of relative bias—most notably when sample sizes are small.


諸外国のデータエディティング及び混淆正規分布モデルによる多変量外れ値検出法についての研究(高橋将宜、選択的エディティング、セレクティブエディティング), Masayoshi Takahashi 2012 National Statistics Center of Japan

諸外国のデータエディティング及び混淆正規分布モデルによる多変量外れ値検出法についての研究(高橋将宜、選択的エディティング、セレクティブエディティング), Masayoshi Takahashi

Masayoshi Takahashi

No abstract provided.


Big Data And The Future, Sherri Rose 2012 Johns Hopkins Bloomberg School of Public Health

Big Data And The Future, Sherri Rose

Sherri Rose

No abstract provided.


A Prior-Free Framework Of Coherent Inference And Its Derivation Of Simple Shrinkage Estimators, David R. Bickel 2012 Ottawa Institute of Systems Biology, Department of Biochemistry, Microbiology, and Immunology, University of Ottawa

A Prior-Free Framework Of Coherent Inference And Its Derivation Of Simple Shrinkage Estimators, David R. Bickel

COBRA Preprint Series

The reasoning behind uses of confidence intervals and p-values in scientific practice may be made coherent by modeling the inferring statistician or scientist as an idealized intelligent agent. With other things equal, such an agent regards a hypothesis coinciding with a confidence interval of a higher confidence level as more certain than a hypothesis coinciding with a confidence interval of a lower confidence level. The agent uses different methods of confidence intervals conditional on what information is available. The coherence requirement means all levels of certainty of hypotheses about the parameter agree with the same distribution of certainty over parameter …


Methods For Shape-Constrained Kernel Density Estimation, Mark A. Wolters 2012 The University of Western Ontario

Methods For Shape-Constrained Kernel Density Estimation, Mark A. Wolters

Electronic Thesis and Dissertation Repository

Nonparametric density estimators are used to estimate an unknown probability density while making minimal assumptions about its functional form. Although the low reliance of nonparametric estimators on modelling assumptions is a benefit, their performance will be improved if auxiliary information about the density's shape is incorporated into the estimate. Auxiliary information can take the form of shape constraints, such as unimodality or symmetry, that the estimate must satisfy. Finding the constrained estimate is usually a difficult optimization problem, however, and a consistent framework for finding estimates across a variety of problems is lacking.

It is proposed to find shape-constrained density …


Targeted Maximum Likelihood Estimation For Dynamic Treatment Regimes In Sequential Randomized Controlled Trials, Paul Chaffee, Mark J. van der Laan 2012 University of California, Berkeley, Division of Biostatistics

Targeted Maximum Likelihood Estimation For Dynamic Treatment Regimes In Sequential Randomized Controlled Trials, Paul Chaffee, Mark J. Van Der Laan

Paul H. Chaffee

Sequential Randomized Controlled Trials (SRCTs) are rapidly becoming essential tools in the search for optimized treatment regimes in ongoing treatment settings. Analyzing data for multiple time-point treatments with a view toward optimal treatment regimes is of interest in many types of afflictions: HIV infection, Attention Deficit Hyperactivity Disorder in children, leukemia, prostate cancer, renal failure, and many others. Methods for analyzing data from SRCTs exist but they are either inefficient or suffer from the drawbacks of estimating equation methodology. We describe an estimation procedure, targeted maximum likelihood estimation (TMLE), which has been fully developed and implemented in point treatment settings, …


Digital Commons powered by bepress