Open Access. Powered by Scholars. Published by Universities.®

Statistics and Probability Commons

Open Access. Powered by Scholars. Published by Universities.®

COBRA

2017

Discipline
Keyword
Publication

Articles 1 - 21 of 21

Full-Text Articles in Statistics and Probability

Using Multilevel Outcomes To Construct And Select Biomarker Combinations For Single-Level Prediction, Allison Meisner, Chirag R. Parikh, Kathleen F. Kerr Oct 2017

Using Multilevel Outcomes To Construct And Select Biomarker Combinations For Single-Level Prediction, Allison Meisner, Chirag R. Parikh, Kathleen F. Kerr

UW Biostatistics Working Paper Series

Biomarker studies may involve a multilevel outcome, such as no, mild, or severe disease. There is often interest in predicting one particular level of the outcome due to its clinical significance. The standard approach to constructing biomarker combinations in this context involves dichotomizing the outcome and using a binary logistic regression model. We assessed whether information can be usefully gained from instead using more sophisticated regression methods. Furthermore, it is often necessary to select among several candidate biomarker combinations. One strategy involves selecting a combination on the basis of its ability to predict the outcome level of interest. We propose …


Constructing A Confidence Interval For The Fraction Who Benefit From Treatment, Using Randomized Trial Data, Emily J. Huang, Ethan X. Fang, Daniel F. Hanley, Michael Rosenblum Oct 2017

Constructing A Confidence Interval For The Fraction Who Benefit From Treatment, Using Randomized Trial Data, Emily J. Huang, Ethan X. Fang, Daniel F. Hanley, Michael Rosenblum

Johns Hopkins University, Dept. of Biostatistics Working Papers

The fraction who benefit from treatment is the proportion of patients whose potential outcome under treatment is better than that under control. Inference on this parameter is challenging since it is only partially identifiable, even in our context of a randomized trial. We propose a new method for constructing a confidence interval for the fraction, when the outcome is ordinal or binary. Our confidence interval procedure is pointwise consistent. It does not require any assumptions about the joint distribution of the potential outcomes, although it has the flexibility to incorporate various user-defined assumptions. Unlike existing confidence interval methods for partially …


Comparison Of Adaptive Randomized Trial Designs For Time-To-Event Outcomes That Expand Versus Restrict Enrollment Criteria, To Test Non-Inferiority, Josh Betz, Jon Arni Steingrimsson, Tianchen Qian, Michael Rosenblum Sep 2017

Comparison Of Adaptive Randomized Trial Designs For Time-To-Event Outcomes That Expand Versus Restrict Enrollment Criteria, To Test Non-Inferiority, Josh Betz, Jon Arni Steingrimsson, Tianchen Qian, Michael Rosenblum

Johns Hopkins University, Dept. of Biostatistics Working Papers

Adaptive enrichment designs involve preplanned rules for modifying patient enrollment criteria based on data accrued in an ongoing trial. These designs may be useful when it is suspected that a subpopulation, e.g., defined by a biomarker or risk score measured at baseline, may benefit more from treatment than the complementary subpopulation. We compare two types of such designs, for the case of two subpopulations that partition the overall population. The first type starts by enrolling the subpopulation where it is suspected the new treatment is most likely to work, and then may expand inclusion criteria if there is early evidence …


Nonparametric Variable Importance Assessment Using Machine Learning Techniques, Brian D. Williamson, Peter B. Gilbert, Noah Simon, Marco Carone Aug 2017

Nonparametric Variable Importance Assessment Using Machine Learning Techniques, Brian D. Williamson, Peter B. Gilbert, Noah Simon, Marco Carone

UW Biostatistics Working Paper Series

In a regression setting, it is often of interest to quantify the importance of various features in predicting the response. Commonly, the variable importance measure used is determined by the regression technique employed. For this reason, practitioners often only resort to one of a few regression techniques for which a variable importance measure is naturally defined. Unfortunately, these regression techniques are often sub-optimal for predicting response. Additionally, because the variable importance measures native to different regression techniques generally have a different interpretation, comparisons across techniques can be difficult. In this work, we study a novel variable importance measure that can …


Combining Biomarkers By Maximizing The True Positive Rate For A Fixed False Positive Rate, Allison Meisner, Marco Carone, Margaret Pepe, Kathleen F. Kerr Jul 2017

Combining Biomarkers By Maximizing The True Positive Rate For A Fixed False Positive Rate, Allison Meisner, Marco Carone, Margaret Pepe, Kathleen F. Kerr

UW Biostatistics Working Paper Series

Biomarkers abound in many areas of clinical research, and often investigators are interested in combining them for diagnosis, prognosis and screening. In many applications, the true positive rate for a biomarker combination at a prespecified, clinically acceptable false positive rate is the most relevant measure of predictive capacity. We propose a distribution-free method for constructing biomarker combinations by maximizing the true positive rate while constraining the false positive rate. Theoretical results demonstrate good operating characteristics for the resulting combination. In simulations, the biomarker combination provided by our method demonstrated improved operating characteristics in a variety of scenarios when compared with …


Developing Biomarker Combinations In Multicenter Studies Via Direct Maximization And Penalization, Allison Meisner, Chirag R. Parikh, Kathleen F. Kerr Jul 2017

Developing Biomarker Combinations In Multicenter Studies Via Direct Maximization And Penalization, Allison Meisner, Chirag R. Parikh, Kathleen F. Kerr

UW Biostatistics Working Paper Series

When biomarker studies involve patients at multiple centers and the goal is to develop biomarker combinations for diagnosis, prognosis, or screening, we consider evaluating the predictive capacity of a given combination with the center-adjusted AUC (aAUC), a summary of conditional performance. Rather than using a general method to construct the biomarker combination, such as logistic regression, we propose estimating the combination by directly maximizing the aAUC. Furthermore, it may be desirable to have a biomarker combination with similar predictive capacity across centers. To that end, we allow for penalization of the variability in center-specific performance. We demonstrate good asymptotic properties …


Evaluation Of Multiple Interventions Using A Stepped Wedge Design, Vivian H. Lyons, Lingyu Li, James Hughes, Ali Rowhani-Rahbar Jun 2017

Evaluation Of Multiple Interventions Using A Stepped Wedge Design, Vivian H. Lyons, Lingyu Li, James Hughes, Ali Rowhani-Rahbar

UW Biostatistics Working Paper Series

Background: Stepped wedge cluster randomized trials are a class of unidirectional crossover studies that have historically been limited to evaluating a single intervention. This design is especially suitable for pragmatic trials where the study feasibility can be improved with a phased introduction of the intervention. We examined variations of stepped wedge designs that would support evaluation of multiple interventions. Methods: We propose four different design variants for implementing a stepped wedge trial with two interventions: concurrent design, supplementation, replacement, and factorial designs. Analyses were conducted comparing the precision of the estimated intervention effects for the different designs. Results: Concurrent, …


Biomarker Combinations For Diagnosis And Prognosis In Multicenter Studies: Principles And Methods, Allison Meisner, Chirag R. Parikh, Kathleen F. Kerr Jun 2017

Biomarker Combinations For Diagnosis And Prognosis In Multicenter Studies: Principles And Methods, Allison Meisner, Chirag R. Parikh, Kathleen F. Kerr

UW Biostatistics Working Paper Series

Many investigators are interested in combining biomarkers to predict an outcome of interest or detect underlying disease. This endeavor is complicated by the fact that many biomarker studies involve data from multiple centers. Depending upon the relationship between center, the biomarkers, and the target of prediction, care must be taken when constructing and evaluating combinations of biomarkers. We introduce a taxonomy to describe the role of center and consider how a biomarker combination should be constructed and evaluated. We show that ignoring center, which is frequently done by clinical researchers, is often not appropriate. The limited statistical literature proposes using …


Optimal, Two Stage, Adaptive Enrichment Designs For Randomized Trials Using Sparse Linear Programming, Michael Rosenblum, Xingyuan Fang, Han Liu Jun 2017

Optimal, Two Stage, Adaptive Enrichment Designs For Randomized Trials Using Sparse Linear Programming, Michael Rosenblum, Xingyuan Fang, Han Liu

Johns Hopkins University, Dept. of Biostatistics Working Papers

Adaptive enrichment designs involve preplanned rules for modifying enrollment criteria based on accruing data in a randomized trial. We focus on designs where the overall population is partitioned into two predefined subpopulations, e.g., based on a biomarker or risk score measured at baseline. The goal is to learn which populations benefit from an experimental treatment. Two critical components of adaptive enrichment designs are the decision rule for modifying enrollment, and the multiple testing procedure. We provide a general method for simultaneously optimizing these components for two stage, adaptive enrichment designs. We minimize the expected sample size under constraints on power …


Estimating Autoantibody Signatures To Detect Autoimmune Disease Patient Subsets, Zhenke Wu, Livia Casciola-Rosen, Ami A. Shah, Antony Rosen, Scott L. Zeger Apr 2017

Estimating Autoantibody Signatures To Detect Autoimmune Disease Patient Subsets, Zhenke Wu, Livia Casciola-Rosen, Ami A. Shah, Antony Rosen, Scott L. Zeger

Johns Hopkins University, Dept. of Biostatistics Working Papers

Autoimmune diseases are characterized by highly specific immune responses against molecules in self-tissues. Different autoimmune diseases are characterized by distinct immune responses, making autoantibodies useful for diagnosis and prediction. In many diseases, the targets of autoantibodies are incompletely defined. Although the technologies for autoantibody discovery have advanced dramatically over the past decade, each of these techniques generates hundreds of possibilities, which are onerous and expensive to validate. We set out to establish a method to greatly simplify autoantibody discovery, using a pre-filtering step to define subgroups with similar specificities based on migration of labeled, immunoprecipitated proteins on sodium dodecyl sulfate …


Evaluation Of Progress Towards The Unaids 90-90-90 Hiv Care Cascade: A Description Of Statistical Methods Used In An Interim Analysis Of The Intervention Communities In The Search Study, Laura Balzer, Joshua Schwab, Mark J. Van Der Laan, Maya L. Petersen Feb 2017

Evaluation Of Progress Towards The Unaids 90-90-90 Hiv Care Cascade: A Description Of Statistical Methods Used In An Interim Analysis Of The Intervention Communities In The Search Study, Laura Balzer, Joshua Schwab, Mark J. Van Der Laan, Maya L. Petersen

U.C. Berkeley Division of Biostatistics Working Paper Series

WHO guidelines call for universal antiretroviral treatment, and UNAIDS has set a global target to virally suppress most HIV-positive individuals. Accurate estimates of population-level coverage at each step of the HIV care cascade (testing, treatment, and viral suppression) are needed to assess the effectiveness of "test and treat" strategies implemented to achieve this goal. The data available to inform such estimates, however, are susceptible to informative missingness: the number of HIV-positive individuals in a population is unknown; individuals tested for HIV may not be representative of those whom a testing intervention fails to reach, and HIV-positive individuals with a viral …


Adaptive Non-Inferiority Margins Under Observable Non-Constancy, Brett S. Hanscom, Deborah J. Donnell, Brian D. Williamson, Jim Hughes Feb 2017

Adaptive Non-Inferiority Margins Under Observable Non-Constancy, Brett S. Hanscom, Deborah J. Donnell, Brian D. Williamson, Jim Hughes

UW Biostatistics Working Paper Series

A central assumption in the design and conduct of non-inferiority trials is that the active-control therapy will have the same degree of effectiveness in the planned non-inferiority trial as it had in the prior placebo-controlled trials used to define the non-inferiority margin. This is referred to as the `constancy' assumption. If the constancy assumption fails, the chosen non-inferiority margin is not valid and the study runs the risk of approving an inferior product or failing to approve a beneficial product. The constancy assumption cannot be validated in a trial without a placebo arm, and it is unlikely ever to be …


It's All About Balance: Propensity Score Matching In The Context Of Complex Survey Data, David Lenis, Trang Q. ;Nguyen, Nian Dong, Elizabeth A. Stuart Feb 2017

It's All About Balance: Propensity Score Matching In The Context Of Complex Survey Data, David Lenis, Trang Q. ;Nguyen, Nian Dong, Elizabeth A. Stuart

Johns Hopkins University, Dept. of Biostatistics Working Papers

Many research studies aim to draw causal inferences using data from large, nationally representative survey samples, and many of these studies use propensity score matching to make those causal inferences as rigorous as possible given the non-experimental nature of the data. However, very few applied studies are careful about incorporating the survey design with the propensity score analysis, which may mean that the results don’t generate population inferences. This may be because few methodological studies examine how to best combine these methods. Furthermore, even fewer of the methodological studies incorporate different non-response mechanisms in their analysis. This study examines methods …


Estimating The Probability Of Clonal Relatedness Of Pairs Of Tumors In Cancer Patients, Audrey Mauguen, Venkatraman E. Seshan, Irina Ostrovnaya, Colin B. Begg Feb 2017

Estimating The Probability Of Clonal Relatedness Of Pairs Of Tumors In Cancer Patients, Audrey Mauguen, Venkatraman E. Seshan, Irina Ostrovnaya, Colin B. Begg

Memorial Sloan-Kettering Cancer Center, Dept. of Epidemiology & Biostatistics Working Paper Series

Next generation sequencing panels are being used increasingly in cancer research to study tumor evolution. A specific statistical challenge is to compare the mutational profiles in different tumors from a patient to determine the strength of evidence that the tumors are clonally related, i.e. derived from a single, founder clonal cell. The presence of identical mutations in each tumor provides evidence of clonal relatedness, although the strength of evidence from a match is related to how commonly the mutation is seen in the tumor type under investigation. This evidence must be weighed against the evidence in favor of independent tumors …


Predicting Future Years Of Life, Health, And Functional Ability: A Healthy Life Calculator For Older Adults, Paula Diehr, Michael Diehr, Alice M. Arnold, Laura Yee, Michelle C. Odden, Calvin H. Hirsch, Stephen Thielke, Bruce Psaty, W Craig Johnson, Jorge Kizer, Anne B. Newman Jan 2017

Predicting Future Years Of Life, Health, And Functional Ability: A Healthy Life Calculator For Older Adults, Paula Diehr, Michael Diehr, Alice M. Arnold, Laura Yee, Michelle C. Odden, Calvin H. Hirsch, Stephen Thielke, Bruce Psaty, W Craig Johnson, Jorge Kizer, Anne B. Newman

UW Biostatistics Working Paper Series

Introduction

Planning for the future would be easier if we knew how long we will live and, more importantly, how many years we will be healthy and able to enjoy it. There are few well-documented aids for predicting our future health. We attempted to meet this need for persons 65 years of age and older.

Methods

Data came from the Cardiovascular Health Study, a large longitudinal study of older adults that began in 1990. Years of life (YOL) were defined by measuring time to death. Years of healthy life (YHL) were defined by an annual question about self-rated health, and …


Quantifying The Totality Of Treatment Effect With Multiple Event-Time Observations In The Presence Of A Terminal Event From A Comparative Clinical Study, Brian Claggett, Lu Tian, Haoda Fu, Scott D. Solomon, L. J. Wei Jan 2017

Quantifying The Totality Of Treatment Effect With Multiple Event-Time Observations In The Presence Of A Terminal Event From A Comparative Clinical Study, Brian Claggett, Lu Tian, Haoda Fu, Scott D. Solomon, L. J. Wei

Harvard University Biostatistics Working Paper Series

To evaluate the totality of one treatment's benefit/risk profile relative to an alternative treatment via a longitudinal comparative clinical study, the timing and occurrence of multiple clinical events are typically collected during the patient's followup. These multiple observations reflect the patient's disease progression/burden over time. The standard practice is to create a composite endpoint from the multiple outcomes, the timing of the occurrence of the first clinical event, to evaluate the treatment via the standard survival analysis techniques. By ignoring all events after the composite outcome, this type of assessment may not be ideal. Various parametric or semi-parametric procedures have …


Tuning Parameter Selection In Cox Proportional Hazards Model With A Diverging Number Of Parameters, Andy Ni, Jianwen Cai Jan 2017

Tuning Parameter Selection In Cox Proportional Hazards Model With A Diverging Number Of Parameters, Andy Ni, Jianwen Cai

Memorial Sloan-Kettering Cancer Center, Dept. of Epidemiology & Biostatistics Working Paper Series

Regularized variable selection is a powerful tool for identifying the true regression model from a large number of candidates by applying penalties to the objective functions. The penalty functions typically involve a tuning parameter that control the complexity of the selected model. The ability of the regularized variable selection methods to identify the true model critically depends on the correct choice of the tuning parameter. In this study we develop a consistent tuning parameter selection method for regularized Cox's proportional hazards model with a diverging number of parameters. The tuning parameter is selected by minimizing the generalized information criterion. We …


Variance Prior Specification For A Basket Trial Design Using Bayesian Hierarchical Modeling, Kristen Cunanan, Alexia Iasonos, Ronglai Shen, Mithat Gonen Jan 2017

Variance Prior Specification For A Basket Trial Design Using Bayesian Hierarchical Modeling, Kristen Cunanan, Alexia Iasonos, Ronglai Shen, Mithat Gonen

Memorial Sloan-Kettering Cancer Center, Dept. of Epidemiology & Biostatistics Working Paper Series

Background: In the era of targeted therapies, clinical trials in oncology are rapidly evolving, wherein patients from multiple diseases are now enrolled and treated according to their genomic mutation(s). In such trials, known as basket trials, the different disease cohorts form the different baskets for inference. Several approaches have been proposed in the literature to efficiently use information from all baskets while simultaneously screening to find individual baskets where the drug works. Most proposed methods are developed in a Bayesian paradigm that requires specifying a prior distribution for a variance parameter, which controls the degree to which information is shared …


Optimized Variable Selection Via Repeated Data Splitting, Marinela Capanu, Colin B. Begg, Mithat Gonen Jan 2017

Optimized Variable Selection Via Repeated Data Splitting, Marinela Capanu, Colin B. Begg, Mithat Gonen

Memorial Sloan-Kettering Cancer Center, Dept. of Epidemiology & Biostatistics Working Paper Series

We introduce a new variable selection procedure that repeatedly splits the data into two sets, one for estimation and one for validation, to obtain an empirically optimized threshold which is then used to screen for variables to include in the final model. Simulation results show that the proposed variable selection technique enjoys superior performance compared to candidate methods, being amongst those with the lowest inclusion of noisy predictors while having the highest power to detect the correct model and being unaffected by correlations among the predictors. We illustrate the methods by applying them to a cohort of patients undergoing hepatectomy …


Mediation Analysis For Censored Survival Data Under An Accelerated Failure Time Model, Isabel Fulcher, Eric J. Tchetgen Tchetgen, Paige Williams Jan 2017

Mediation Analysis For Censored Survival Data Under An Accelerated Failure Time Model, Isabel Fulcher, Eric J. Tchetgen Tchetgen, Paige Williams

Harvard University Biostatistics Working Paper Series

Recent advances in causal mediation analysis have formalized conditions for estimating direct and indirect effects in various contexts. These approaches have been extended to a number of models for survival outcomes including accelerated failure time (AFT) models which are widely used in a broad range of health applications given their intuitive interpretation. In this setting, it has been suggested that under standard assumptions, the “difference” and “product” methods produce equivalent estimates of the indirect effect of exposure on the survival outcome. We formally show that these two methods may produce substantially different estimates in the presence of censoring or truncation, …


Studying The Optimal Scheduling For Controlling Prostate Cancer Under Intermittent Androgen Suppression, Sunil K. Dhar, Hans R. Chaudhry, Bruce G. Bukiet, Zhiming Ji, Nan Gao, Thomas W. Findley Jan 2017

Studying The Optimal Scheduling For Controlling Prostate Cancer Under Intermittent Androgen Suppression, Sunil K. Dhar, Hans R. Chaudhry, Bruce G. Bukiet, Zhiming Ji, Nan Gao, Thomas W. Findley

Harvard University Biostatistics Working Paper Series

This retrospective study shows that the majority of patients’ correlations between PSA and Testosterone during the on-treatment period is at least 0.90. Model-based duration calculations to control PSA levels during off-treatment are provided. There are two pairs of models. In one pair, the Generalized Linear Model and Mixed Model are both used to analyze the variability of PSA at the individual patient level by using the variable “Patient ID” as a repeated measure. In the second pair, Patient ID is not used as a repeated measure but additional baseline variables are included to analyze the variability of PSA.