Estimation Of The Treatment Effect With Bayesian Adjustment For Covariates, 2020 University of Kentucky
Estimation Of The Treatment Effect With Bayesian Adjustment For Covariates, Li Xu
Theses and Dissertations--Statistics
The Bayesian adjustment for confounding (BAC) is a Bayesian model averaging method to select and adjust for confounding factors when evaluating the average causal effect of an exposure on a certain outcome. We extend the BAC method to time-to-event outcomes. Specifically, the posterior distribution of the exposure effect on a time-to-event outcome is calculated as a weighted average of posterior distributions from a number of candidate proportional hazards models, weighing each model by its ability to adjust for confounding factors. The Bayesian Information Criterion based on the partial likelihood is used to compare different models and approximate the Bayes factor. …
Phenotype Extraction: Estimation And Biometrical Genetic Analysis Of Individual Dynamics, 2020 Virginia Commonwealth University
Phenotype Extraction: Estimation And Biometrical Genetic Analysis Of Individual Dynamics, Kevin L. Mckee
Theses and Dissertations
Within-person data can exhibit a virtually limitless variety of statistical patterns, but it can be difficult to distinguish meaningful features from statistical artifacts. Studies of complex traits have previously used genetic signals like twin-based heritability to distinguish between the two. This dissertation is a collection of studies applying state-space modeling to conceptualize and estimate novel phenotypic constructs for use in psychiatric research and further biometrical genetic analysis. The aims are to: (1) relate control theoretic concepts to health-related phenotypes; (2) design statistical models that formally define those phenotypes; (3) estimate individual phenotypic values from time series data; (4) consider hierarchical …
Aggregate Loss Model With Poisson-Tweedie Loss Frequency, 2020 Wilfrid Laurier University
Aggregate Loss Model With Poisson-Tweedie Loss Frequency, Si Chen
Theses and Dissertations (Comprehensive)
The aggregate loss model has applications in various areas such as financial risk management and actuarial science. The aggregate loss is the summation of all random losses occurred in a period, and it is governed by both the loss severity and the loss frequency. While the impact of the loss severity on aggregate loss is well studied, less focus is paid on the influence of loss frequency on aggregate loss, which motivates our study. In this thesis, we enrich the aggregate loss framework by introducing the Poisson-Tweedie distribution as a candidate for modelling loss frequency, prove the closedness of Poisson-Tweedie …
How Machine Learning And Probability Concepts Can Improve Nba Player Evaluation, 2020 Claremont Colleges
How Machine Learning And Probability Concepts Can Improve Nba Player Evaluation, Harrison Miller
CMC Senior Theses
In this paper I will be breaking down a scholarly article, written by Sameer K. Deshpande and Shane T. Jensen, that proposed a new method to evaluate NBA players. The NBA is the highest level professional basketball league in America and stands for the National Basketball Association. They proposed to build a model that would result in how NBA players impact their teams chances of winning a game, using machine learning and probability concepts. I preface that by diving into these concepts and their mathematical backgrounds. These concepts include building a linear model using ordinary least squares method, the bias …
Identifying Customer Churn In After-Market Operations Using Machine Learning Algorithms, 2019 Southern Methodist University
Identifying Customer Churn In After-Market Operations Using Machine Learning Algorithms, Vitaly Briker, Richard Farrow, William Trevino, Brent Allen
SMU Data Science Review
This paper presents a comparative study on machine learning methods as they are applied to product associations, future purchase predictions, and predictions of customer churn in aftermarket operations. Association rules are used help to identify patterns across products and find correlations in customer purchase behaviour. Studying customer behaviour as it pertains to Recency, Frequency, and Monetary Value (RFM) helps inform customer segmentation and identifies customers with propensity to churn. Lastly, Flowserve’s customer purchase history enables the establishment of churn thresholds for each customer group and assists in constructing a model to predict future churners. The aim of this model is …
Personalized Detection Of Anxiety Provoking News Events Using Semantic Network Analysis, 2019 Southern Methodist University
Personalized Detection Of Anxiety Provoking News Events Using Semantic Network Analysis, Jacquelyn Cheun Phd, Luay Dajani, Quentin B. Thomas
SMU Data Science Review
In the age of hyper-connectivity, 24/7 news cycles, and instant news alerts via social media, mental health researchers don't have a way to automatically detect news content which is associated with triggering anxiety or depression in mental health patients. Using the Associated Press news wire, a semantic network was built with 1,056 news articles containing over 500,000 connections across multiple topics to provide a personalized algorithm which detects problematic news content for a given reader. We make use of Semantic Network Analysis to surface the relationship between news article text and anxiety in readers who struggle with mental health disorders. …
Ordinal Hyperplane Loss, 2019 Kennesaw State University
Ordinal Hyperplane Loss, Bob Vanderheyden
Doctor of Data Science and Analytics Dissertations
This research presents the development of a new framework for analyzing ordered class data, commonly called “ordinal class” data. The focus of the work is the development of classifiers (predictive models) that predict classes from available data. Ratings scales, medical classification scales, socio-economic scales, meaningful groupings of continuous data, facial emotional intensity and facial age estimation are examples of ordinal data for which data scientists may be asked to develop predictive classifiers. It is possible to treat ordinal classification like any other classification problem that has more than two classes. Specifying a model with this strategy does not fully utilize …
Evaluation Of Modern Missing Data Handling Methods For Coefficient Alpha, 2019 University of Nebraska - Lincoln
Evaluation Of Modern Missing Data Handling Methods For Coefficient Alpha, Katerina Matysova
College of Education and Human Sciences: Dissertations, Theses, and Student Research
When assessing a certain characteristic or trait using a multiple item measure, quality of that measure can be assessed by examining the reliability. To avoid multiple time points, reliability can be represented by internal consistency, which is most commonly calculated using Cronbach’s coefficient alpha. Almost every time human participants are involved in research, there is missing data involved. Missing data means that even though complete data were expected to be collected, some data are missing. Missing data can follow different patterns as well as be the result of different mechanisms. One traditional way to deal with missing data is listwise …
Seasonal Time Series Models With Application To Weather And Lake Level Data, 2019 Missouri State University
Seasonal Time Series Models With Application To Weather And Lake Level Data, Mengqing Qin
MSU Graduate Theses
This work studies seasonal time series models with application to lake level and weather data. The thesis includes related time series concepts, integrated autoregressive moving average models (abbreviated as ARIMA), parameter estimation, model diagnostics, and forecasting. The studied time series models are applied to the data of daily lake level in Beaver Lake (1988-2017) and the data of daily maximum temperature in New York Central Park (1870-2017). Due to seasonality of the data, three different approaches are proposed to the modeling: regression method, functional ARIMA method and multiplicative seasonal ARIMA method. The forecasted values of the year 2018 are compared …
Conditional Survival Analysis For Concrete Bridge Decks, 2019 University of Wisconsin-Milwaukee
Conditional Survival Analysis For Concrete Bridge Decks, Azam Nabizadeh, Habib Tabatabai, Mohammad A. Tabatabai
Civil and Environmental Engineering Faculty Articles
Bridge decks are a significant factor in the deterioration of bridges, and substantially affect long-term bridge maintenance decisions. In this study, conditional survival (reliability) analysis techniques are applied to bridge decks to evaluate the age at the end of service life using the National Bridge Inventory records. As bridge decks age, the probability of survival and the expected service life would change. The additional knowledge gained from the fact that a bridge deck has already survived a specific number of years alters (increases) the original probability of survival at subsequent years based on the conditional probability theory. The conditional expected …
Habitat Associations And Reproduction Of Fishes On The Northwestern Gulf Of Mexico Shelf Edge, 2019 Louisiana State University and Agricultural and Mechanical College
Habitat Associations And Reproduction Of Fishes On The Northwestern Gulf Of Mexico Shelf Edge, Elizabeth Marie Keller
LSU Doctoral Dissertations
Several of the northwestern Gulf of Mexico (GOM) shelf-edge banks provide critical hard bottom habitat for coral and fish communities, supporting a wide diversity of ecologically and economically important species. These sites may be fish aggregation and spawning sites and provide important habitat for fish growth and reproduction. Already designated as habitat areas of particular concern, many of these banks are also under consideration for inclusion in the expansion of the Flower Garden Banks National Marine Sanctuary. This project aimed to gain a more comprehensive understanding of the communities and fish species on shelf-edge banks by way of gonad histology, …
Inferring A Consensus Problem List Using Penalized Multistage Models For Ordered Data, 2019 The University Of Michigan
Inferring A Consensus Problem List Using Penalized Multistage Models For Ordered Data, Philip S. Boonstra, John C. Krauss
The University of Michigan Department of Biostatistics Working Paper Series
A patient's medical problem list describes his or her current health status and aids in the coordination and transfer of care between providers, among other things. Because a problem list is generated once and then subsequently modified or updated, what is not usually observable is the provider-effect. That is, to what extent does a patient's problem in the electronic medical record actually reflect a consensus communication of that patient's current health status? To that end, we report on and analyze a unique interview-based design in which multiple medical providers independently generate problem lists for each of three patient case abstracts …
An Agent-Based Modeling Approach For Predicting The Behavior Of Bighead Carp (Hypophthalmichthys Nobilis) Under The Influence Of Acoustic Deterrence, 2019 Valparaiso University
An Agent-Based Modeling Approach For Predicting The Behavior Of Bighead Carp (Hypophthalmichthys Nobilis) Under The Influence Of Acoustic Deterrence, Joey Gaudy, Craig Garzella
Annual Symposium on Biomathematics and Ecology Education and Research
No abstract provided.
Integrating Mathematics And Biology In The Classroom: A Compendium Of Case Studies And Labs, 2019 University of North Carolina at Asheville
Integrating Mathematics And Biology In The Classroom: A Compendium Of Case Studies And Labs, Becky Sanft, Anne Walter
Annual Symposium on Biomathematics and Ecology Education and Research
No abstract provided.
Classification Of Coronary Artery Disease In Non-Diabetic Patients Using Artificial Neural Networks, 2019 Illinois State University
Classification Of Coronary Artery Disease In Non-Diabetic Patients Using Artificial Neural Networks, Demond Handley
Annual Symposium on Biomathematics and Ecology Education and Research
No abstract provided.
Statistical Modeling And Characterization Of Induced Seismicity Within The Western Canada Sedimentary Basin, 2019 The University of Western Ontario
Statistical Modeling And Characterization Of Induced Seismicity Within The Western Canada Sedimentary Basin, Sid Kothari
Electronic Thesis and Dissertation Repository
In western Canada, there has been an increase in seismic activity linked to anthropogenic energy-related operations including conventional hydrocarbon production, wastewater fluid injection and more recently hydraulic fracturing (HF). Statistical modeling and characterization of the space, time and magnitude distributions of the seismicity clusters is vital for a better understanding of induced earthquake processes and development of predictive models. In this work, a statistical analysis of the seismicity in the Western Canada Sedimentary Basin was performed across past and present time periods by utilizing a compiled earthquake catalogue for Alberta and eastern British Columbia. Specifically, the frequency-magnitude statistics were analyzed …
Statistical L-Moment And L-Moment Ratio Estimation And Their Applicability In Network Analysis, 2019 Air Force Institute of Technology
Statistical L-Moment And L-Moment Ratio Estimation And Their Applicability In Network Analysis, Timothy S. Anderson
Theses and Dissertations
This research centers on finding the statistical moments, network measures, and statistical tests that are most sensitive to various node degradations for the Barabási-Albert, Erdös-Rényi, and Watts-Strogratz network models. Thirty-five different graph structures were simulated for each of the random graph generation algorithms, and sensitivity analysis was undertaken on three different network measures: degree, betweenness, and closeness. In an effort to find the statistical moments that are the most sensitive to degradation within each network, four traditional moments: mean, variance, skewness, and kurtosis as well as three non-traditional moments: L-variance, L-skewness, and L-kurtosis were examined. Each of these moments were …
Sample Size Requirements And Considerations For Models To Assess Human-Machine System Performance, 2019 Air Force Institute of Technology
Sample Size Requirements And Considerations For Models To Assess Human-Machine System Performance, Jennifer S. G. Lopez
Theses and Dissertations
Hierarchical Linear Models (HLMs), also known as multi-level models, are an extension of multiple regression analysis and can aid in the understanding of human and machine workloads of a system. These models allow for prediction and testing in systems with hierarchies of two or more levels. The complex interrelated variability of these multi-level models exists in operational settings, such as the Air Force Distributed Common Ground System Full Motion Video (AF DCGS FMV) community which is composed of individuals (Level-1), groups (Level-2), units (Level-3), and organizations (Level-4). Through the development of sample size requirements and considerations for multi-level models, this …
Semi-Supervised Regression With Generative Adversarial Networks Using Minimal Labeled Data, 2019 The Graduate Center, City University of New York
Semi-Supervised Regression With Generative Adversarial Networks Using Minimal Labeled Data, Greg Olmschenk
Dissertations, Theses, and Capstone Projects
This work studies the generalization of semi-supervised generative adversarial networks (GANs) to regression tasks. A novel feature layer contrasting optimization function, in conjunction with a feature matching optimization, allows the adversarial network to learn from unannotated data and thereby reduce the number of labels required to train a predictive network. An analysis of simulated training conditions is performed to explore the capabilities and limitations of the method. In concert with the semi-supervised regression GANs, an improved label topology and upsampling technique for multi-target regression tasks are shown to reduce data requirements. Improvements are demonstrated on a wide variety of vision …
Machine Learning In Support Of Electric Distribution Asset Failure Prediction, 2019 Southern Methodist University
Machine Learning In Support Of Electric Distribution Asset Failure Prediction, Robert D. Flamenbaum, Thomas Pompo, Christopher Havenstein, Jade Thiemsuwan
SMU Data Science Review
In this paper, we present novel approaches to predicting as- set failure in the electric distribution system. Failures in overhead power lines and their associated equipment in particular, pose significant finan- cial and environmental threats to electric utilities. Electric device failure furthermore poses a burden on customers and can pose serious risk to life and livelihood. Working with asset data acquired from an electric utility in Southern California, and incorporating environmental and geospatial data from around the region, we applied a Random Forest methodology to predict which overhead distribution lines are most vulnerable to fail- ure. Our results provide evidence …