Open Access. Powered by Scholars. Published by Universities.®

Physical Sciences and Mathematics Commons

Open Access. Powered by Scholars. Published by Universities.®

COBRA

COBRA Preprint Series

Series

Discipline
Keyword
Publication Year

Articles 1 - 30 of 87

Full-Text Articles in Physical Sciences and Mathematics

An Improved Bayesian Pick-The-Winner (Ibpw) Design For Randomized Phase Ii Clinical Trials, Wanni Lei, Maosen Peng, Xi K. Zhou May 2024

An Improved Bayesian Pick-The-Winner (Ibpw) Design For Randomized Phase Ii Clinical Trials, Wanni Lei, Maosen Peng, Xi K. Zhou

COBRA Preprint Series

Phase II clinical trials play a pivotal role in drug development by screening a large number of drug candidates to identify those with promising preliminary efficacy for phase III testing. Trial designs that enable efficient decision-making with small sample sizes and early futility stopping while controlling for type I and II errors in hypothesis testing, such as Simon’s two-stage design, are preferred. Randomized multi-arm trials are increasingly used in phase II settings to overcome the limitations associated with using historical controls as the reference. However, how to effectively balance efficiency and accurate decision-making continues to be an important research topic. …


A Simple And Robust Alternative To Bland-Altman Method Of Assessing Clinical Agreement, Abhaya Indrayan Prof Jan 2022

A Simple And Robust Alternative To Bland-Altman Method Of Assessing Clinical Agreement, Abhaya Indrayan Prof

COBRA Preprint Series

Clinical agreement between two quantitative measurements on a group of subjects is generally assessed with the help of the Bland-Altman (B-A) limits. These limits only describe the dispersion of disagreements in 95% cases and do not measure the degree of agreement. The interpretation regarding the presence or absence of agreement by this method is based on whether B-A limits are within the pre-specified externally determined clinical tolerance limits. Thus, clinical tolerance limits are necessary for this method. We argue in this communication that the direct use of clinical tolerance limits for assessing agreement without the B-A limits is more effective …


Unified Methods For Feature Selection In Large-Scale Genomic Studies With Censored Survival Outcomes, Lauren Spirko-Burns, Karthik Devarajan Mar 2019

Unified Methods For Feature Selection In Large-Scale Genomic Studies With Censored Survival Outcomes, Lauren Spirko-Burns, Karthik Devarajan

COBRA Preprint Series

One of the major goals in large-scale genomic studies is to identify genes with a prognostic impact on time-to-event outcomes which provide insight into the disease's process. With rapid developments in high-throughput genomic technologies in the past two decades, the scientific community is able to monitor the expression levels of tens of thousands of genes and proteins resulting in enormous data sets where the number of genomic features is far greater than the number of subjects. Methods based on univariate Cox regression are often used to select genomic features related to survival outcome; however, the Cox model assumes proportional hazards …


A Spline-Assisted Semiparametric Approach To Nonparametric Measurement Error Models, Fei Jiang, Yanyuan Ma Mar 2018

A Spline-Assisted Semiparametric Approach To Nonparametric Measurement Error Models, Fei Jiang, Yanyuan Ma

COBRA Preprint Series

Nonparametric estimation of the probability density function of a random variable measured with error is considered to be a difficult problem, in the sense that depending on the measurement error prop- erty, the estimation rate can be as slow as the logarithm of the sample size. Likewise, nonparametric estimation of the regression function with errors in the covariate suffers the same possibly slow rate. The traditional methods for both problems are based on deconvolution, where the slow convergence rate is caused by the quick convergence to zero of the Fourier transform of the measurement error density, which, unfortunately, appears in …


Hpcnmf: A High-Performance Toolbox For Non-Negative Matrix Factorization, Karthik Devarajan, Guoli Wang Feb 2016

Hpcnmf: A High-Performance Toolbox For Non-Negative Matrix Factorization, Karthik Devarajan, Guoli Wang

COBRA Preprint Series

Non-negative matrix factorization (NMF) is a widely used machine learning algorithm for dimension reduction of large-scale data. It has found successful applications in a variety of fields such as computational biology, neuroscience, natural language processing, information retrieval, image processing and speech recognition. In bioinformatics, for example, it has been used to extract patterns and profiles from genomic and text-mining data as well as in protein sequence and structure analysis. While the scientific performance of NMF is very promising in dealing with high dimensional data sets and complex data structures, its computational cost is high and sometimes could be critical for …


Semi-Parametrics Dose Finding Methods, Matthieu Clertant, John O'Quigley Jan 2016

Semi-Parametrics Dose Finding Methods, Matthieu Clertant, John O'Quigley

COBRA Preprint Series

We describe a new class of dose finding methods to be used in early phase clinical trials. Under some added parametric conditions the class reduces to the family of continual reassessment method (CRM) designs. Under some relaxation of the underlying structure the method is equivalent to the CCD, mTPI or BOIN classes of designs. These latter designs are non-parametric in nature whereas the CRM class can be viewed as being strongly parametric. The proposed class is characterized as being semi-parametric since it corresponds to CRM with a nuisance parameter. Performance is good, matching that of the CRM class and improving …


Statistical Handling Of Medical Data - An Ethical Perspective, Ajay Kumar Bansal Dr Dec 2015

Statistical Handling Of Medical Data - An Ethical Perspective, Ajay Kumar Bansal Dr

COBRA Preprint Series

Medical Science is a delicate subject and the clinical data generated from the medical trials must be reliable and of good quality. Not only the quality of generated data is important, but the management is also crucial and is to be handled very carefully. In this paper, the ethical aspect of statistical handling of such data is discussed.

Every profession has some set of norms to follow to achieve its objectives. These norms are called professional ethics which shows the essence of human behaviour. Same way, the field of medical research is expected to follow ethical norms, to obtain reliable …


A Simple Method To Estimate The Time-Dependent Roc Curve Under Right Censoring, Liang Li, Bo Hu, Tom Greene Sep 2015

A Simple Method To Estimate The Time-Dependent Roc Curve Under Right Censoring, Liang Li, Bo Hu, Tom Greene

COBRA Preprint Series

The time-dependent Receiver Operating Characteristic (ROC) curve is often used to study the diagnostic accuracy of a single continuous biomarker, measured at baseline, on the onset of a disease condition when the disease onset may occur at different times during the follow-up and hence may be right censored. Due to censoring, the true disease onset status prior to the pre-specified time horizon may be unknown on some patients, which causes difficulty in calculating the time-dependent sensitivity and specificity. We study a simple method that adjusts for censoring by weighting the censored data by the conditional probability of disease onset prior …


Distance Correlation Measures Applied To Analyze Relation Between Variables In Liver Cirrhosis Marker Data, Atanu Bhattacharjee Dr. May 2015

Distance Correlation Measures Applied To Analyze Relation Between Variables In Liver Cirrhosis Marker Data, Atanu Bhattacharjee Dr.

COBRA Preprint Series

Distance Correlation is another newer choice to compute the relation between variables. However, the Bayesian counterpart of Distance Correlation is not established. In this paper, Bayesian counterpart of Distance Correlation is pro- posed. Proposed method is illustrated with Liver Chirrhosis Marker data. The relevant studies information about relation between AST and ALT is used to formulate the prior information for Bayesian computation. The Distance Correlation between AST and ALT (both are liver performance marker) is computed with 0.44. The credible interval is observed with (0.41, 0.46).Bayesian counter- part to compute Distance correlation is simple and handy.


Methods For Exploring Treatment Effect Heterogeneity In Subgroup Analysis: An Application To Global Clinical Trials, I. Manjula Schou, Ian C. Marschner Jun 2014

Methods For Exploring Treatment Effect Heterogeneity In Subgroup Analysis: An Application To Global Clinical Trials, I. Manjula Schou, Ian C. Marschner

COBRA Preprint Series

Multi-country randomised clinical trials (MRCTs) are common in the medical literature and their interpretation has been the subject of extensive recent discussion. In many MRCTs, an evaluation of treatment effect homogeneity across countries or regions is conducted. Subgroup analysis principles require a significant test of interaction in order to claim heterogeneity of treatment effect across subgroups, such as countries in a MRCT. As clinical trials are typically underpowered for tests of interaction, overly optimistic expectations of treatment effect homogeneity can lead researchers, regulators and other stakeholders to over-interpret apparent differences between subgroups even when heterogeneity tests are insignificant. In this …


Variable Selection For Zero-Inflated And Overdispersed Data With Application To Health Care Demand In Germany, Zhu Wang, Shuangge Ma, Ching-Yun Wang May 2014

Variable Selection For Zero-Inflated And Overdispersed Data With Application To Health Care Demand In Germany, Zhu Wang, Shuangge Ma, Ching-Yun Wang

COBRA Preprint Series

In health services and outcome research, count outcomes are frequently encountered and often have a large proportion of zeros. The zero-inflated negative binomial (ZINB) regression model has important applications for this type of data. With many possible candidate risk factors, this paper proposes new variable selection methods for the ZINB model. We consider maximum likelihood function plus a penalty including the least absolute shrinkage and selection operator (LASSO), smoothly clipped absolute deviation (SCAD) and minimax concave penalty (MCP). An EM (expectation-maximization) algorithm is proposed for estimating the model parameters and conducting variable selection simultaneously. This algorithm consists of estimating penalized …


Pre-Maceration, Saignée And Temperature Affect Daily Evolution Of Pigment Extraction During Vinification, Ottorino L. Pantani, Federico M. Stefanini, Irene Lozzi, Luca Calamai, Alessandra Biondi Bartolini, Stefano Di Blasi Apr 2014

Pre-Maceration, Saignée And Temperature Affect Daily Evolution Of Pigment Extraction During Vinification, Ottorino L. Pantani, Federico M. Stefanini, Irene Lozzi, Luca Calamai, Alessandra Biondi Bartolini, Stefano Di Blasi

COBRA Preprint Series

Consumer demand for intensely coloured wines necessitates the systematic testing of pigment extraction in Sangiovese, a cultivar poor in easily extractable anthocyanins. Pre-fermentation (absent, cold soak pre-fermentation at 5 °C, cryomaceration by liquid N2 addition), temperature (20 or 30 °C), and saignée were compared during vinification (800 kg). Concentrations of anthocyanins, non-anthocyanic flavonoids and SO2-resistant pigments were recorded daily. A semiparametric Bayesian model permitted the kinetic description and the comparison of sigmoidal- and exponential-like curves. In total anthocyanins, saignée at 30 °C yielded a significant gain, later lost at drawing off; cryomaceration had little effect and cold …


Bayesian Model Averaging:- An Application In Cancer Clinical Trial, Atanu Bhattacharjee Feb 2014

Bayesian Model Averaging:- An Application In Cancer Clinical Trial, Atanu Bhattacharjee

COBRA Preprint Series

Data driven conclusion is mostly accepted approach in any medical research problem. In case of limited knowledge of deep idea about supportive data on the problem, automatic digging of the variable plays important role for insight view of the study. Bayesian model averaging can be considered for automatics variable selection. It can be used as an alternative of stepwise regression method. The aim of this paper is to show the application of Bayesian modeling averaging in medical research particularly in cancer trial. Method is illustrated on Bone marrow transplant data. It can be recommended that BMA can be used frequently …


Computational Model For Survey And Trend Analysis Of Patients With Endometriosis : A Decision Aid Tool For Ebm, Salvo Reina, Vito Reina, Franco Ameglio, Mauro Costa, Alessandro Fasciani Feb 2014

Computational Model For Survey And Trend Analysis Of Patients With Endometriosis : A Decision Aid Tool For Ebm, Salvo Reina, Vito Reina, Franco Ameglio, Mauro Costa, Alessandro Fasciani

COBRA Preprint Series

Endometriosis is increasingly collecting worldwide attention due to its medical complexity and social impact. The European community has identified this as a “social disease”. A large amount of information comes from scientists, yet several aspects of this pathology and staging criteria need to be clearly defined on a suitable number of individuals. In fact, available studies on endometriosis are not easily comparable due to a lack of standardized criteria to collect patients’ informations and scarce definitions of symptoms. Currently, only retrospective surgical stadiation is used to measure pathology intensity, while the Evidence Based Medicine (EBM) requires shareable methods and correct …


Regression Trees For Longitudinal Data, Madan Gopal Kundu, Jaroslaw Harezlak Sep 2013

Regression Trees For Longitudinal Data, Madan Gopal Kundu, Jaroslaw Harezlak

COBRA Preprint Series

Often when a longitudinal change is studied in a population of interest we find that changes over time are heterogeneous (in terms of time and/or covariates' effect) and a traditional linear mixed effect model [Laird and Ware, 1982] on the entire population assuming common parametric form for covariates and time may not be applicable to the entire population. This is usually the case in studies when there are many possible predictors influencing the response trajectory. For example, Raudenbush [2001] used depression as an example to argue that it is incorrect to assume that all the people in a given population …


Augmentation Of Propensity Scores For Medical Records-Based Research, Mikel Aickin Jun 2013

Augmentation Of Propensity Scores For Medical Records-Based Research, Mikel Aickin

COBRA Preprint Series

Therapeutic research based on electronic medical records suffers from the possibility of various kinds of confounding. Over the past 30 years, propensity scores have increasingly been used to try to reduce this possibility. In this article a gap is identified in the propensity score methodology, and it is proposed to augment traditional treatment-propensity scores with outcome-propensity scores, thereby removing all other aspects of common causes from the analysis of treatment effects.


A Bayesian Regression Tree Approach To Identify The Effect Of Nanoparticles Properties On Toxicity Profiles, Cecile Low-Kam, Haiyuan Zhang, Zhaoxia Ji, Tian Xia, Jeffrey I. Zinc, Andre Nel, Donatello Telesca Mar 2013

A Bayesian Regression Tree Approach To Identify The Effect Of Nanoparticles Properties On Toxicity Profiles, Cecile Low-Kam, Haiyuan Zhang, Zhaoxia Ji, Tian Xia, Jeffrey I. Zinc, Andre Nel, Donatello Telesca

COBRA Preprint Series

We introduce a Bayesian multiple regression tree model to characterize relationships between physico-chemical properties of nanoparticles and their in-vitro toxicity over multiple doses and times of exposure. Unlike conventional models that rely on data summaries, our model solves the low sample size issue and avoids arbitrary loss of information by combining all measurements from a general exposure experiment across doses, times of exposure, and replicates. The proposed technique integrates Bayesian trees for modeling threshold effects and interactions, and penalized B-splines for dose and time-response surfaces smoothing. The resulting posterior distribution is sampled via a Markov Chain Monte Carlo algorithm. This …


Relating Nanoparticle Properties To Biological Outcomes In Exposure Escalation Experiments, Trina Patel, Cecile Low-Kam, Zhaoxia Ji, Haiyuan Zhang, Tian Xia, Andre E. Nel, Jeffrey I. Zinc, Donatello Telesca Dec 2012

Relating Nanoparticle Properties To Biological Outcomes In Exposure Escalation Experiments, Trina Patel, Cecile Low-Kam, Zhaoxia Ji, Haiyuan Zhang, Tian Xia, Andre E. Nel, Jeffrey I. Zinc, Donatello Telesca

COBRA Preprint Series

A fundamental goal in nano-toxicology is that of identifying particle physical and chemical properties, which are likely to explain biological hazard. The first line of screening for potentially adverse outcomes often consists of exposure escalation experiments, involving the exposure of micro-organisms or cell lines to a battery of nanomaterials. We discuss a modeling strategy, that relates the outcome of an exposure escalation experiment to nanoparticle properties. Our approach makes use of a hierarchical decision process, where we jointly identify particles that initiate adverse biological outcomes and explain the probability of this event in terms of the particle physico-chemical descriptors. The …


Pls-Rog: Partial Least Squares With Rank Order Of Groups, Hiroyuki Yamamoto Oct 2012

Pls-Rog: Partial Least Squares With Rank Order Of Groups, Hiroyuki Yamamoto

COBRA Preprint Series

Partial least squares (PLS), which is an unsupervised dimensionality reduction method, has been widely used in metabolomics. PLS can separate score depend on groups in a low dimensional subspace. However, this cannot use the information about rank order of groups. This information is often provided in which concentration of administered drugs to animals is gradually varies. In this study, we proposed partial least squares for rank order of groups (PLS-ROG). PLS-ROG can consider both separation and rank order of groups.


Statistical Hypothesis Test Of Factor Loading In Principal Component Analysis And Its Application To Metabolite Set Enrichment Analysis, Hiroyuki Yamamoto, Tamaki Fujimori, Hajime Sato, Gen Ishikawa, Kenjiro Kami, Yoshiaki Ohashi Oct 2012

Statistical Hypothesis Test Of Factor Loading In Principal Component Analysis And Its Application To Metabolite Set Enrichment Analysis, Hiroyuki Yamamoto, Tamaki Fujimori, Hajime Sato, Gen Ishikawa, Kenjiro Kami, Yoshiaki Ohashi

COBRA Preprint Series

Principal component analysis (PCA) has been widely used to visualize high-dimensional metabolomic data in a two- or three-dimensional subspace. In metabolomics, some metabolites (e.g. top 10 metabolites) have been subjectively selected when using factor loading in PCA, and biological inferences for these metabolites are made. However, this approach is possible to lead biased biological inferences because these metabolites are not objectively selected by statistical criterion. We proposed a statistical procedure to pick up metabolites by statistical hypothesis test of factor loading in PCA and make biological inferences by metabolite set enrichment analysis (MSEA) for these significant metabolites. This procedure depends …


Quantifying Alternative Splicing From Paired-End Rna-Sequencing Data, David Rossell, Camille Stephan-Otto Attolini, Manuel Kroiss, Almond Stöcker Sep 2012

Quantifying Alternative Splicing From Paired-End Rna-Sequencing Data, David Rossell, Camille Stephan-Otto Attolini, Manuel Kroiss, Almond Stöcker

COBRA Preprint Series

RNA-sequencing has revolutionized biomedical research and, in particular, our ability to study gene alternative splicing. The problem has important implications for human health, as alternative splicing is involved in malfunctions at the cellular level and multiple diseases. However, the high-dimensional nature of the data and the existence of experimental biases pose serious data analysis challenges. We find that the standard data summaries used to study alternative splicing are severely limited, as they ignore a substantial amount of valuable information. Current data analysis methods are based on such summaries and are hence sub-optimal. Further, they have limited flexibility in accounting for …


Robust Estimation Of Pure/Natural Direct Effects With Mediator Measurement Error, Eric J. Tchetgen Tchetgen, Sheng Hsuan Lin Sep 2012

Robust Estimation Of Pure/Natural Direct Effects With Mediator Measurement Error, Eric J. Tchetgen Tchetgen, Sheng Hsuan Lin

COBRA Preprint Series

Recent developments in causal mediation analysis have offered new notions of direct and indirect effects, that formalize more traditional and informal notions of mediation analysis emanating primarily from the social sciences. The pure or natural direct effect of Robins-Greenland-Pearl quantifies the causal effect of an exposure that is not mediated by a variable on the causal pathway to the outcome, and combines with the natural indirect effect to produce the total causal effect of the exposure. Sufficient conditions for identification of natural direct effects were previously given, that assume certain independencies about potential outcomes, and a rich literature on estimation …


A Prior-Free Framework Of Coherent Inference And Its Derivation Of Simple Shrinkage Estimators, David R. Bickel Jun 2012

A Prior-Free Framework Of Coherent Inference And Its Derivation Of Simple Shrinkage Estimators, David R. Bickel

COBRA Preprint Series

The reasoning behind uses of confidence intervals and p-values in scientific practice may be made coherent by modeling the inferring statistician or scientist as an idealized intelligent agent. With other things equal, such an agent regards a hypothesis coinciding with a confidence interval of a higher confidence level as more certain than a hypothesis coinciding with a confidence interval of a lower confidence level. The agent uses different methods of confidence intervals conditional on what information is available. The coherence requirement means all levels of certainty of hypotheses about the parameter agree with the same distribution of certainty over parameter …


On Identification Of Natural Direct Effects When A Confounder Of The Mediator Is Directly Affected By Exposure, Eric J. Tchetgen Tchetgen, Tyler J. Vanderweele Jun 2012

On Identification Of Natural Direct Effects When A Confounder Of The Mediator Is Directly Affected By Exposure, Eric J. Tchetgen Tchetgen, Tyler J. Vanderweele

COBRA Preprint Series

Natural direct and indirect effects formalize traditional notions of mediation analysis into a rigorous causal framework and have recently received considerable attention in epidemiology and in the social sciences. Sufficient conditions for identification of natural direct effects were formulated by Judea Pearl under a nonparametric structural equations model, which assumes certain independencies between potential outcomes. A common situation in epidemiology is that a confounder of the mediator is affected by the exposure, in which case, natural direct effects fail to be nonparametrically identified without additional assumptions, even under Pearl's nonparametric structural equations model. In this paper, the authors show that …


Why Odds Ratio Estimates Of Gwas Are Almost Always Close To 1.0, Yutaka Yasui May 2012

Why Odds Ratio Estimates Of Gwas Are Almost Always Close To 1.0, Yutaka Yasui

COBRA Preprint Series

“Missing heritability” in genome-wide association studies (GWAS) refers to the seeming inability for GWAS data to capture the great majority of genetic causes of a disease in comparison to the known degree of heritability for the disease, in spite of GWAS’ genome-wide measures of genetic variations. This paper presents a simple mathematical explanation for this phenomenon, assuming that the heritability information exists in GWAS data. Specifically, it focuses on the fact that the great majority of association measures (in the form of odds ratios) from GWAS are consistently close to the value that indicates no association, explains why this occurs, …


Differential Patterns Of Interaction And Gaussian Graphical Models, Masanao Yajima, Donatello Telesca, Yuan Ji, Peter Muller Apr 2012

Differential Patterns Of Interaction And Gaussian Graphical Models, Masanao Yajima, Donatello Telesca, Yuan Ji, Peter Muller

COBRA Preprint Series

We propose a methodological framework to assess heterogeneous patterns of association amongst components of a random vector expressed as a Gaussian directed acyclic graph. The proposed framework is likely to be useful when primary interest focuses on potential contrasts characterizing the association structure between known subgroups of a given sample. We provide inferential frameworks as well as an efficient computational algorithm to fit such a model and illustrate its validity through a simulation. We apply the model to Reverse Phase Protein Array data on Acute Myeloid Leukemia patients to show the contrast of association structure between refractory patients and relapsed …


Hierarchical Rank Aggregation With Applications To Nanotoxicology, Trina Patel, Donatello Telesca, Robert Rallo, Saji George, Xia Tian, Nel Andre Mar 2012

Hierarchical Rank Aggregation With Applications To Nanotoxicology, Trina Patel, Donatello Telesca, Robert Rallo, Saji George, Xia Tian, Nel Andre

COBRA Preprint Series

The development of high throughput screening (HTS) assays in the field of nanotoxicology provide new opportunities for the hazard assessment and ranking of engineered nanomaterials (ENM). It is often necessary to rank lists of materials based on multiple risk assessment parameters, often aggregated across several measures of toxicity and possibly spanning an array of experimental platforms. Bayesian models coupled with the optimization of loss functions have been shown to provide an effective framework for conducting inference on ranks. In this article we present various loss function based ranking approaches for comparing ENM within experiments and toxicity parameters. Additionally, we propose …


Robustness Of Measures Of Interaction To Unmeasured Confounding, Eric J. Tchetgen Tchetgen, Tyler J. Vanderweele Mar 2012

Robustness Of Measures Of Interaction To Unmeasured Confounding, Eric J. Tchetgen Tchetgen, Tyler J. Vanderweele

COBRA Preprint Series

In this paper, we study the impact of unmeasured confounding on inference about a two-way interaction in a mean regression model with identity, log or logit link function. Necessary and sufficient conditions are established for a two-way interaction to be nonparametrically identified from the observed data, despite unmeasured confounding for the factors defining the interaction. A lung cancer data application illustrates the results.


Toxicity Profiling Of Engineered Nanomaterials Via Multivariate Dose Response Surface Modeling, Trina Patel, Donatello Telesca, Saji George, Andre Nel Dec 2011

Toxicity Profiling Of Engineered Nanomaterials Via Multivariate Dose Response Surface Modeling, Trina Patel, Donatello Telesca, Saji George, Andre Nel

COBRA Preprint Series

New generation in-vitro high throughput screening (HTS) assays for the assessment of engineered nanomaterials provide an opportunity to learn how these particles interact at the cellular level, particularly in relation to injury pathways. These types of assays are often characterized by small sample sizes, high measurement error and high dimensionality as multiple cytotoxicity outcomes are measured across an array of doses and durations of exposure. In this article we propose a probability model for toxicity profiling of engineered nanomaterials. A hierarchical framework is used to account for the multivariate nature of the data by modeling dependence between outcomes and thereby …


Modeling Criminal Careers As Departures From A Unimodal Population Age-Crime Curve: The Case Of Marijuana Use, Donatello Telesca, Elena Erosheva, Derek Kreager, Ross Matsueda Dec 2011

Modeling Criminal Careers As Departures From A Unimodal Population Age-Crime Curve: The Case Of Marijuana Use, Donatello Telesca, Elena Erosheva, Derek Kreager, Ross Matsueda

COBRA Preprint Series

A major aim of longitudinal analyses of life course data is to describe the within- and between-individual variability in a behavioral outcome, such as crime. Statistical analyses of such data typically draw on mixture and mixed-effects growth models. In this work, we present a functional analytic point of view and develop an alternative method that models individual crime trajectories as departures from a population age-crime curve. Drawing on empirical and theoretical claims in criminology, we assume a unimodal population age-crime curve and allow individual expected crime trajectories to differ by their levels of offending and patterns of temporal misalignment. We …