Open Access. Powered by Scholars. Published by Universities.®

Statistical Methodology Commons

Open Access. Powered by Scholars. Published by Universities.®

1,114 Full-Text Articles 1,507 Authors 603,045 Downloads 125 Institutions

All Articles in Statistical Methodology

Faceted Search

1,114 full-text articles. Page 36 of 38.

Curriculum Vitae, Tatiyana V. Apanasovich 2010 Thomas Jefferson University

Curriculum Vitae, Tatiyana V. Apanasovich

Tatiyana V Apanasovich

No abstract provided.


Multilevel Functional Principal Component Analysis For High-Dimensional Data, Vadim Zipunnikov, Brian Caffo, Ciprian Crainiceanu, David M. Yousem, Christos Davatzikos, Brian S. Schwartz 2010 Johns Hopkins Bloomberg School of Public Health, Department of Biostatistics

Multilevel Functional Principal Component Analysis For High-Dimensional Data, Vadim Zipunnikov, Brian Caffo, Ciprian Crainiceanu, David M. Yousem, Christos Davatzikos, Brian S. Schwartz

Johns Hopkins University, Dept. of Biostatistics Working Papers

We propose fast and scalable statistical methods for the analysis of hundreds or thousands of high dimensional vectors observed at multiple visits. The proposed inferential methods avoid the difficult task of loading the entire data set at once in the computer memory and use sequential access to data. This allows deployment of our methodology on low-resource computers where computations can be done in minutes on extremely large data sets. Our methods are motivated by and applied to a study where hundreds of subjects were scanned using Magnetic Resonance Imaging (MRI) at two visits roughly five years apart. The original data …


Landmark Prediction Of Survival, Layla Parast, Tianxi Cai 2010 Harvard School of Public Health

Landmark Prediction Of Survival, Layla Parast, Tianxi Cai

Harvard University Biostatistics Working Paper Series

No abstract provided.


Longitudinal Penalized Functional Regression, Jeff Goldsmith, Ciprian M. Crainiceanu, Brian Caffo, Daniel Reich 2010 Johns Hopkins Bloomberg School of Public Health, Department of Biostatistics

Longitudinal Penalized Functional Regression, Jeff Goldsmith, Ciprian M. Crainiceanu, Brian Caffo, Daniel Reich

Johns Hopkins University, Dept. of Biostatistics Working Papers

We propose a new regression model and inferential tools for the case when both the outcome and the functional exposures are observed at multiple visits. This data structure is new but increasingly present in applications where functions or images are recorded at multiple times. This raises new inferential challenges that cannot be addressed with current methods and software. Our proposed model generalizes the Generalized Linear Mixed Effects Model (GLMM) by adding functional predictors. Smoothness of the functional coefficients is ensured using roughness penalties estimated by Restricted Maximum Likelihood (REML) in a corresponding mixed effects model. This method is computationally feasible …


On Two-Stage Hypothesis Testing Procedures Via Asymptotically Independent Statistics, James Dai, Charles Kooperberg, Michael L. LeBlanc, Ross Prentice 2010 FHCRC

On Two-Stage Hypothesis Testing Procedures Via Asymptotically Independent Statistics, James Dai, Charles Kooperberg, Michael L. Leblanc, Ross Prentice

UW Biostatistics Working Paper Series

Kooperberg and LeBlanc (2008) proposed a two-stage testing procedure to screen for significant interactions in genome-wide association (GWA) studies by a soft threshold on marginal associations (MA), though its theoretical properties and generalization have not been elaborated. In this article, we discuss conditions that are required to achieve strong control of the Family-Wise Error Rate (FWER) by such procedures for low or high-dimensional hypothesis testing. We provide proof of asymptotic independence of marginal association statistics and interaction statistics in linear regression, logistic regression, and Cox proportional hazard models in a randomized clinical trial (RCT) with a rare event. In case-control …


Stratifying Subjects For Treatment Selection With Censored Event Time Data From A Comparative Study, Lihui Zhao, Tianxi Cai, Lu Tian, Hajime Uno, Scott D. Solomon, L. J. Wei 2010 Harvard University

Stratifying Subjects For Treatment Selection With Censored Event Time Data From A Comparative Study, Lihui Zhao, Tianxi Cai, Lu Tian, Hajime Uno, Scott D. Solomon, L. J. Wei

Harvard University Biostatistics Working Paper Series

No abstract provided.


Comparison Of Non-Surgical And Surgical Endodontic Retreatment: A Systematic Review, Robert Corr 2010 Loma Linda University

Comparison Of Non-Surgical And Surgical Endodontic Retreatment: A Systematic Review, Robert Corr

Loma Linda University Electronic Theses, Dissertations & Projects

Primary root canal therapy has been shown to be a predictable procedure with a high degree of success 1-4, however failures can occur after treatment. Treatment options for the preservation of teeth that have had previous endodontic treatment but demonstrate persistent disease include non-surgical (orthograde) or surgical (retrograde) endodontic retreatment, assuming the tooth is restorable, periodontally sound, and the patient desires to retain the tooth. The purpose of this study was to conduct a systematic review of the current available evidence to compare the clinical and radiographic outcomes of nonsurgical with those of surgical endodontic retreatment. Methodology began with …


On Two-Stage Hypothesis Testing Procedures Via Asymptotically Independent Statistics, James Y. Dai, Charles Kooperberg, Michael LeBlanc, Ross L. Prentice 2010 Fred Hutchinson Cancer Research Center

On Two-Stage Hypothesis Testing Procedures Via Asymptotically Independent Statistics, James Y. Dai, Charles Kooperberg, Michael Leblanc, Ross L. Prentice

UW Biostatistics Working Paper Series

Kooperberg08 proposed a two-stage testing procedure to screen for significant interactions in genome-wide association (GWA) studies by a soft threshold on marginal associations (MA), though its theoretical properties and generalization have not been elaborated. In this article, we discuss conditions that are required to achieve strong control of the Family-Wise Error Rate (FWER) by such procedures for low or high-dimensional hypothesis testing. We provide proof of asymptotic independence of marginal association statistics and interaction statistics in linear regression, logistic regression, and Cox proportional hazard models in a randomized clinical trial (RCT) with a rare event. In case-control studies nested within …


A Perturbation Method For Inference On Regularized Regression Estimates, Jessica Minnier, Lu Tian, Tianxi Cai 2010 Harvard University

A Perturbation Method For Inference On Regularized Regression Estimates, Jessica Minnier, Lu Tian, Tianxi Cai

Harvard University Biostatistics Working Paper Series

No abstract provided.


Mixture Of Factor Analyzers With Information Criteria And The Genetic Algorithm, Esra Turan 2010 University of Tennessee, Knoxville

Mixture Of Factor Analyzers With Information Criteria And The Genetic Algorithm, Esra Turan

Doctoral Dissertations

In this dissertation, we have developed and combined several statistical techniques in Bayesian factor analysis (BAYFA) and mixture of factor analyzers (MFA) to overcome the shortcoming of these existing methods. Information Criteria are brought into the context of the BAYFA model as a decision rule for choosing the number of factors m along with the Press and Shigemasu method, Gibbs Sampling and Iterated Conditional Modes deterministic optimization. Because of sensitivity of BAYFA on the prior information of the factor pattern structure, the prior factor pattern structure is learned directly from the given sample observations data adaptively using Sparse Root algorithm. …


Estimating Confidence Intervals For Eigenvalues In Exploratory Factor Analysis, Ross Larsen, Russell Warne 2010 Brigham Young University - Provo

Estimating Confidence Intervals For Eigenvalues In Exploratory Factor Analysis, Ross Larsen, Russell Warne

Russell T Warne

Exploratory factor analysis (EFA) has become a common procedure in educational and psychological research. In the course of performing an EFA, researchers often base the decision of how many factors to retain on the eigenvalues for the factors. However, many researchers do not realize that eigenvalues, like all sample statistics, are subject to sampling error, which means that confidence intervals (CIs) can be estimated for each eigenvalue. In the present article, we demonstrate two methods of estimating CIs for eigenvalues: one based on the mathematical properties of the central limit theorem, and the other based on bootstrapping. References to appropriate …


Principled Sure Independence Screening For Cox Models With Ultra-High-Dimensional Covariates, Sihai Dave Zhao, Yi Li 2010 Harvard School of Public Health and Dana Farber Cancer Institute

Principled Sure Independence Screening For Cox Models With Ultra-High-Dimensional Covariates, Sihai Dave Zhao, Yi Li

Harvard University Biostatistics Working Paper Series

No abstract provided.


Data Mining As Applied To The Socialsciences, Walter Morris 2010 Dyson College of Arts and Sciences, Pace University

Data Mining As Applied To The Socialsciences, Walter Morris

Cornerstone 3 Reports : Interdisciplinary Informatics

No abstract provided.


Optimizing Randomized Trial Designs To Distinguish Which Subpopulations Benefit From Treatment, Michael Rosenblum, Mark J. van der Laan 2010 Johns Hopkins University

Optimizing Randomized Trial Designs To Distinguish Which Subpopulations Benefit From Treatment, Michael Rosenblum, Mark J. Van Der Laan

U.C. Berkeley Division of Biostatistics Working Paper Series

It is a challenge to evaluate experimental treatments where it is suspected that the treatment effect may only be strong for certain subpopulations, such as those having a high initial severity of disease, or those having a particular gene variant. Standard randomized controlled trials can have low power in such situations. They also are not optimized to distinguish which subpopulations benefit from a treatment. With the goal of overcoming these limitations, we consider randomized trial designs in which the criteria for patient enrollment may be changed, in a preplanned manner, based on interim analyses. Since such designs allow data-dependent changes …


The Strength Of Statistical Evidence For Composite Hypotheses: Inference To The Best Explanation, David R. Bickel 2010 Ottawa Institute of Systems Biology, Department of Biochemistry, Microbiology, and Immunology, Department of Mathematics and Statistics

The Strength Of Statistical Evidence For Composite Hypotheses: Inference To The Best Explanation, David R. Bickel

COBRA Preprint Series

A general function to quantify the weight of evidence in a sample of data for one hypothesis over another is derived from the law of likelihood and from a statistical formalization of inference to the best explanation. For a fixed parameter of interest, the resulting weight of evidence that favors one composite hypothesis over another is the likelihood ratio using the parameter value consistent with each hypothesis that maximizes the likelihood function over the parameter of interest. Since the weight of evidence is generally only known up to a nuisance parameter, it is approximated by replacing the likelihood function with …


Model-Robust Regression And A Bayesian `Sandwich' Estimator, Adam A. Szpiro, Kenneth M. Rice, Thomas Lumley 2010 University of Washington

Model-Robust Regression And A Bayesian `Sandwich' Estimator, Adam A. Szpiro, Kenneth M. Rice, Thomas Lumley

UW Biostatistics Working Paper Series

The published version of this paper in Annals of Applied Statistics (Vol. 4, No. 4 (2010), 2099–2113) is available from the journal web site at http://dx.doi.org/10.1214/10-AOAS362.

We present a new Bayesian approach to model-robust linear regression that leads to uncertainty estimates with the same robustness properties as the Huber-White sandwich estimator. The sandwich estimator is known to provide asymptotically correct frequentist inference, even when standard modeling assumptions such as linearity and homoscedasticity in the data-generating mechanism are violated. Our derivation provides a compelling Bayesian justification for using this simple and popular tool, and it also clarifies what is being estimated …


Asymptotic Properties Of The Sequential Empirical Roc And Ppv Curves, Joseph S. Koopmeiners, Ziding Feng 2010 University of Washington & Fred Hutchinson Cancer Research Center

Asymptotic Properties Of The Sequential Empirical Roc And Ppv Curves, Joseph S. Koopmeiners, Ziding Feng

UW Biostatistics Working Paper Series

The receiver operating characteristic (ROC) curve, the positive predictive value (PPV) curve and the negative predictive value (NPV) curve are three common measures of performance for a diagnostic biomarker. The independent increments covariance structure assumption is common in the group sequential study design literature. Showing that summary measures of the ROC, PPV and NPV curves have an independent increments covariance structure will provide the theoretical foundation for designing group sequential diagnostic biomarker studies. The ROC, PPV and NPV curves are often estimated empirically to avoid assumptions about the distributional form of the biomarkers. In this paper we derive asymptotic theory …


Estimating Causal Effects In Trials Involving Multi-Treatment Arms Subject To Non-Compliance: A Bayesian Frame-Work, Qi Long, Roderick J. Little, Xihong Lin 2010 Emory University

Estimating Causal Effects In Trials Involving Multi-Treatment Arms Subject To Non-Compliance: A Bayesian Frame-Work, Qi Long, Roderick J. Little, Xihong Lin

Harvard University Biostatistics Working Paper Series

No abstract provided.


Super Learner In Prediction, Eric C. Polley, Mark J. van der Laan 2010 Division of Biostatistics, University of California, Berkeley

Super Learner In Prediction, Eric C. Polley, Mark J. Van Der Laan

U.C. Berkeley Division of Biostatistics Working Paper Series

Super learning is a general loss based learning method that has been proposed and analyzed theoretically in van der Laan et al. (2007). In this article we consider super learning for prediction. The super learner is a prediction method designed to find the optimal combination of a collection of prediction algorithms. The super learner algorithm finds the combination of algorithms minimizing the cross-validated risk. The super learner framework is built on the theory of cross-validation and allows for a general class of prediction algorithms to be considered for the ensemble. Due to the previously established oracle results for the cross-validation …


A New Screening Methodology For Mixture Experiments, Maria Weese 2010 University of Tennessee - Knoxville

A New Screening Methodology For Mixture Experiments, Maria Weese

Doctoral Dissertations

Many materials we use in daily life are comprised of a mixture; plastics, gasoline, food, medicine, etc. Mixture experiments, where factors are proportions of components and the response depends only on the relative proportions of the components, are an integral part of product development and improvement. However, when the number of components is large and there are complex constraints, experimentation can be a daunting task. We study screening methods in a mixture setting using the framework of the Cox mixture model [1]. We exploit the easy interpretation of the parameters in the Cox mixture model and develop methods for screening …


Digital Commons powered by bepress