Instrumental Variable Analyses: Exploiting Natural Randomness To Understand Causal Mechanisms, 2012 University of Michigan
Instrumental Variable Analyses: Exploiting Natural Randomness To Understand Causal Mechanisms, Theodore Iwashyna, Edward Kennedy
Edward H. Kennedy
Instrumental variable analysis is a technique commonly used in the social sciences to provide evidence that a treatment causes an outcome, as contrasted with evidence that a treatment is merely associated with differences in an outcome. To extract such strong evidence from observational data, instrumental variable analysis exploits situations where some degree of randomness affects how patients are selected for a treatment. An instrumental variable is a characteristic of the world that leads some people to be more likely to get the specific treatment we want to study but does not otherwise change thosepatients’ outcomes. This seminar explains, in nonmathematical …
Theory And Methods For Gini Coefficients Partitioned By Quantile Range, 2012 Fordham University
Theory And Methods For Gini Coefficients Partitioned By Quantile Range, Chaitra Nagaraja
Chaitra H Nagaraja
The Gini coefficient is frequently used to measure inequality in populations. However, it is possible that inequality levels may change over time differently for disparate subgroups which cannot be detected with population-level estimates only. Therefore, it may be informative to examine inequality separately for these segments. The case where the population is split into two segments based on non-overlapping quantile ranges is examined. Asymptotic theory is derived and practical methods to estimate standard errors and construct confidence intervals using resampling methods are developed. An application to per capita income across census tracts using American Community Survey data is considered.
A Comparison Of Periodic Autoregressive And Dynamic Factor Models In Intraday Energy Demand Forecasting, 2012 Melbourne Business School
A Comparison Of Periodic Autoregressive And Dynamic Factor Models In Intraday Energy Demand Forecasting, Thomas Mestekemper, Goeran Kauermann, Michael Smith
Michael Stanley Smith
We suggest a new approach for forecasting energy demand at an intraday resolution. Demand in each intraday period is modeled using semiparametric regression smoothing to account for calendar and weather components. Residual serial dependence is captured by one of two multivariate stationary time series models, with dimension equal to the number of intraday periods. These are a periodic autoregression and a dynamic factor model. We show the benefits of our approach in the forecasting of district heating demand in a steam network in Germany and aggregate electricity demand in the state of Victoria, Australia. In both studies, accounting for weather …
An Overview Of Targeted Maximum Likelihood Estimation, 2012 Harvard School of Public Health
An Overview Of Targeted Maximum Likelihood Estimation, Susan Gruber
Susan Gruber
These slides provide an introduction to targeted maximum likelihood estimation in a point treatment setting.
Bayesian Approaches To Copula Modelling, 2012 Melbourne Business School
Bayesian Approaches To Copula Modelling, Michael S. Smith
Michael Stanley Smith
Copula models have become one of the most widely used tools in the applied modelling of multivariate data. Similarly, Bayesian methods are increasingly used to obtain efficient likelihood-based inference. However, to date, there has been only limited use of Bayesian approaches in the formulation and estimation of copula models. This article aims to address this shortcoming in two ways. First, to introduce copula models and aspects of copula theory that are especially relevant for a Bayesian analysis. Second, to outline Bayesian approaches to formulating and estimating copula models, and their advantages over alternative methods. Copulas covered include Archimedean, copulas constructed …
Bayesian Methods For Expression-Based Integration, 2012 Texas A&M University
Bayesian Methods For Expression-Based Integration, Elizabeth M. Jennings, Jeffrey S. Morris, Raymond J. Carroll, Ganiraju C. Manyam, Veera Baladandayuthapani
Jeffrey S. Morris
We propose methods to integrate data across several genomic platforms using a hierarchical Bayesian analysis framework that incorporates the biological relationships among the platforms to identify genes whose expression is related to clinical outcomes in cancer. This integrated approach combines information across all platforms, leading to increased statistical power in finding these predictive genes, and further provides mechanistic information about the manner in which the gene affects the outcome. We demonstrate the advantages of the shrinkage estimation used by this approach through a simulation, and finally, we apply our method to a Glioblastoma Multiforme dataset and identify several genes potentially …
Quantifying Temporal Correlations: A Test-Retest Evaluation Of Functional Connectivity In Resting-State Fmri, 2012 University of California - San Diego
Quantifying Temporal Correlations: A Test-Retest Evaluation Of Functional Connectivity In Resting-State Fmri, Mark Fiecas, Hernando Ombao, Dan Van Lunen, Richard Baumgartner, Alexandre Coimbra, Dai Feng
Mark Fiecas
There have been many interpretations of functional connectivity and proposed measures of temporal correlations between BOLD signals across different brain areas. These interpretations yield from many studies on functional connectivity using resting-state fMRI data that have emerged in recent years. However, not all of these studies used the same metrics for quantifying the temporal correlations between brain regions. In this paper, we use a public-domain test–retest resting-state fMRI data set to perform a systematic investigation of the stability of the metrics that are often used in resting-state functional connectivity (FC) studies. The fMRI data set was collected across three different …
Methods For Evaluating Prediction Performance Of Biomarkers And Tests, 2012 Fred Hutchinson Cancer Research Center
Methods For Evaluating Prediction Performance Of Biomarkers And Tests, Margaret S. Pepe Phd, Holly Janes Phd
Margaret S Pepe PhD
This chapter describes and critiques methods for evaluating the performance of markers to predict risk of a current or future clinical outcome. We consider three criteria that are important for evaluating a risk model: calibration, benefit for decision making and accurate classification. We also describe and discuss a variety of summary measures in common use for quantifying predictive information such as the area under the ROC curve and R-squared. The roles and problems with recently proposed risk reclassification approaches are discussed in detail.
Group Testing Regression Models, 2012 University of Nebraska-Lincoln
Group Testing Regression Models, Boan Zhang
Department of Statistics: Dissertations, Theses, and Student Work
Group testing, where groups of individual specimens are composited to test for the presence or absence of a disease (or some other binary characteristic), is a procedure commonly used to reduce the costs of screening a large number of individuals. Statistical research in group testing has traditionally focused on a homogeneous population, where individuals are assumed to have the same probability of having a disease. However, individuals often have different risks of positivity, so recent research has examined regression models that allow for heterogeneity among individuals within the population. This dissertation focuses on two problems involving group testing regression models. …
Obtaining Critical Values For Test Of Markov Regime Switching, 2012 University of California, Santa Barbara
Obtaining Critical Values For Test Of Markov Regime Switching, Douglas G. Steigerwald, Valerie Bostwick
Douglas G. Steigerwald
For Markov regime-switching models, testing for the possible presence of more than one regime requires the use of a non-standard test statistic. Carter and Steigerwald (forthcoming, Journal of Econometric Methods) derive in detail the analytic steps needed to implement the test ofMarkov regime-switching proposed by Cho and White (2007, Econometrica). We summarize the implementation steps and address the computational issues that arise. A new command to compute regime-switching critical values, rscv, is introduced and presented in the context of empirical research.
Quest For Continuous Improvement: Gathering Feedback And Data Through Multiple Methods To Evaluate And Improve A Library’S Discovery Tool, 2012 University of Nevada, Las Vegas
Quest For Continuous Improvement: Gathering Feedback And Data Through Multiple Methods To Evaluate And Improve A Library’S Discovery Tool, Jeanne M. Brown
Library Faculty Presentations
Summon at UNLV
- Implemented fall 2011: a web-scale discovery tool
- Expectations for Summon
- Continuous Summon Improvement (CSI)Group
The environment
- User changes
- Library changes
- Vendor changes
- Product changes
- Complex information environment
- Change + complexity = need to assess using multiple streams of feedback
A Doubling Technique For The Power Method Transformations, 2012 University of Texas at Arlington
A Doubling Technique For The Power Method Transformations, Mohan D. Pant, Todd C. Headrick
Mohan Dev Pant
Power method polynomials are used for simulating non-normal distributions with specified product moments or L-moments. The power method is capable of producing distributions with extreme values of skew (L-skew) and kurtosis (L-kurtosis). However, these distributions can be extremely peaked and thus not representative of real-world data. To obviate this problem, two families of distributions are introduced based on a doubling technique with symmetric standard normal and logistic power method distributions. The primary focus of the methodology is in the context of L-moment theory. As such, L-moment based systems of equations are derived for simulating univariate and multivariate non-normal distributions with …
Finding A Better Confidence Interval For A Single Regression Changepoint Using Different Bootstrap Confidence Interval Procedures, 2012 Georgia Southern University
Finding A Better Confidence Interval For A Single Regression Changepoint Using Different Bootstrap Confidence Interval Procedures, Bodhipaksha Thilakarathne
Electronic Theses and Dissertations
Recently a number of papers have been published in the area of regression changepoints but there is not much literature concerning confidence intervals for regression changepoints. The purpose of this paper is to find a better bootstrap confidence interval for a single regression changepoint. ("Better" confidence interval means having a minimum length and coverage probability which is close to a chosen significance level). Several methods will be used to find bootstrap confidence intervals. Among those methods a better confidence interval will be presented.
Adventures In Library Salary Surveys, 2012 University of Vermont
Adventures In Library Salary Surveys, Scott L. Schaffer
UVM Libraries Conference Day
Salary surveys are an important tool for the library community and the administrators and boards responsible for the oversight of libraries. However, such assessments must be constructed and analyzed with great care. The Vermont Library Association Personnel Committee has conducted three salary surveys over the past several years, one focusing on academic libraries and two on public libraries. Significant issues have included confidentiality, participation rate, definitions, length and difficulty of questions, collection of data, and representativeness. Suggestions and lessons learned will be shared.
An L-Moment-Based Analog For The Schmeiser-Deutsch Class Of Distributions, 2012 Southern Illinois University Carbondale
An L-Moment-Based Analog For The Schmeiser-Deutsch Class Of Distributions, Todd C. Headrick, Mohan D. Pant
Mohan Dev Pant
This paper characterizes the conventional moment-based Schmeiser-Deutsch (S-D) class of distributions through the method of L-moments. The system can be used in a variety of settings such as simulation or modeling various processes. A procedure is also described for simulating S-D distributions with specified L-moments and L-correlations. The Monte Carlo results presented in this study indicate that the estimates of L-skew, L-kurtosis, and L-correlation associated with the S-D class of distributions are substantially superior to their corresponding conventional product-moment estimators in terms of relative bias—most notably when sample sizes are small.
諸外国のデータエディティング及び混淆正規分布モデルによる多変量外れ値検出法についての研究(高橋将宜、選択的エディティング、セレクティブエディティング), 2012 National Statistics Center of Japan
諸外国のデータエディティング及び混淆正規分布モデルによる多変量外れ値検出法についての研究(高橋将宜、選択的エディティング、セレクティブエディティング), Masayoshi Takahashi
Masayoshi Takahashi
No abstract provided.
Big Data And The Future, 2012 Johns Hopkins Bloomberg School of Public Health
A Prior-Free Framework Of Coherent Inference And Its Derivation Of Simple Shrinkage Estimators, 2012 Ottawa Institute of Systems Biology, Department of Biochemistry, Microbiology, and Immunology, University of Ottawa
A Prior-Free Framework Of Coherent Inference And Its Derivation Of Simple Shrinkage Estimators, David R. Bickel
COBRA Preprint Series
The reasoning behind uses of confidence intervals and p-values in scientific practice may be made coherent by modeling the inferring statistician or scientist as an idealized intelligent agent. With other things equal, such an agent regards a hypothesis coinciding with a confidence interval of a higher confidence level as more certain than a hypothesis coinciding with a confidence interval of a lower confidence level. The agent uses different methods of confidence intervals conditional on what information is available. The coherence requirement means all levels of certainty of hypotheses about the parameter agree with the same distribution of certainty over parameter …
Methods For Shape-Constrained Kernel Density Estimation, 2012 The University of Western Ontario
Methods For Shape-Constrained Kernel Density Estimation, Mark A. Wolters
Electronic Thesis and Dissertation Repository
Nonparametric density estimators are used to estimate an unknown probability density while making minimal assumptions about its functional form. Although the low reliance of nonparametric estimators on modelling assumptions is a benefit, their performance will be improved if auxiliary information about the density's shape is incorporated into the estimate. Auxiliary information can take the form of shape constraints, such as unimodality or symmetry, that the estimate must satisfy. Finding the constrained estimate is usually a difficult optimization problem, however, and a consistent framework for finding estimates across a variety of problems is lacking.
It is proposed to find shape-constrained density …
Targeted Maximum Likelihood Estimation For Dynamic Treatment Regimes In Sequential Randomized Controlled Trials, 2012 University of California, Berkeley, Division of Biostatistics
Targeted Maximum Likelihood Estimation For Dynamic Treatment Regimes In Sequential Randomized Controlled Trials, Paul Chaffee, Mark J. Van Der Laan
Paul H. Chaffee
Sequential Randomized Controlled Trials (SRCTs) are rapidly becoming essential tools in the search for optimized treatment regimes in ongoing treatment settings. Analyzing data for multiple time-point treatments with a view toward optimal treatment regimes is of interest in many types of afflictions: HIV infection, Attention Deficit Hyperactivity Disorder in children, leukemia, prostate cancer, renal failure, and many others. Methods for analyzing data from SRCTs exist but they are either inefficient or suffer from the drawbacks of estimating equation methodology. We describe an estimation procedure, targeted maximum likelihood estimation (TMLE), which has been fully developed and implemented in point treatment settings, …