Open Access. Powered by Scholars. Published by Universities.®

Digital Commons Network

Open Access. Powered by Scholars. Published by Universities.®

Articles 1 - 23 of 23

Full-Text Articles in Entire DC Network

Life Data Analysis Of Repairable Systems: A Case Study On Brigham Young University Media Rooms, Stephen Oluaku Manortey Dec 2006

Life Data Analysis Of Repairable Systems: A Case Study On Brigham Young University Media Rooms, Stephen Oluaku Manortey

Theses and Dissertations

It is an undisputable fact that most systems, upon consistence usage are bound to fail in the performance of their intended functions at a point in time. When this occurs, various strategies are set in place to restore them back to a satisfactory performance. This may include replacing the failed component with a new one, swapping parts, resetting adjustable parts to mention but a few. Any such system is referred to as a repairable system. There is the need to study these systems and use statistical models to predict their failing time and be able to set modalities in place …


A Comparison Of Microarray Analyses: A Mixed Models Approach Versus The Significance Analysis Of Microarrays, Nathan Wallace Stephens Nov 2006

A Comparison Of Microarray Analyses: A Mixed Models Approach Versus The Significance Analysis Of Microarrays, Nathan Wallace Stephens

Theses and Dissertations

DNA microarrays are a relatively new technology for assessing the expression levels of thousands of genes simultaneously. Researchers hope to find genes that are differentially expressed by hybridizing cDNA from known treatment sources with various genes spotted on the microarrays. The large number of tests involved in analyzing microarrays has raised new questions in multiple testing. Several approaches for identifying differentially expressed genes have been proposed. This paper considers two: (1) a mixed models approach, and (2) the Signiffcance Analysis of Microarrays.


A Logistic Regression Analysis Of Utah Colleges Exit Poll Response Rates Using Sas Software, Clint W. Stevenson Oct 2006

A Logistic Regression Analysis Of Utah Colleges Exit Poll Response Rates Using Sas Software, Clint W. Stevenson

Theses and Dissertations

In this study I examine voter response at an interview level using a dataset of 7562 voter contacts (including responses and nonresponses) in the 2004 Utah Colleges Exit Poll. In 2004, 4908 of the 7562 voters approached responded to the exit poll for an overall response rate of 65 percent. Logistic regression is used to estimate factors that contribute to a success or failure of each interview attempt. This logistic regression model uses interviewer characteristics, voter characteristics (both respondents and nonrespondents), and exogenous factors as independent variables. Voter characteristics such as race, gender, and age are strongly associated with response. …


Theory Of Effectiveness Measurement, Richard K. Bullock Sep 2006

Theory Of Effectiveness Measurement, Richard K. Bullock

Theses and Dissertations

Effectiveness measures provide decision makers feedback on the impact of deliberate actions and affect critical issues such as allocation of scarce resources, as well as whether to maintain or change existing strategy. Currently, however, there is no formal foundation for formulating effectiveness measures. This research presents a new framework for effectiveness measurement from both a theoretical and practical view. First, accepted effects-based principles, as well as fundamental measurement concepts are combined into a general, domain independent, effectiveness measurement methodology. This is accomplished by defining effectiveness measurement as the difference, or conceptual distance from a given system state to some reference …


Statistical Approach To Background Subtraction For Production Of High-Quality Silhouettes For Human Gait Recognition, Jennifer J. Samler Sep 2006

Statistical Approach To Background Subtraction For Production Of High-Quality Silhouettes For Human Gait Recognition, Jennifer J. Samler

Theses and Dissertations

This thesis uses a background subtraction to produce high-quality silhouettes for use in human identification by human gait recognition, an identification method which does not require contact with an individual and which can be done from a distance. A statistical method which reduces the noise level is employed resulting in cleaner silhouettes which facilitate identification. The thesis starts with gathering video data of individuals walking normally across a background scene. From there the video is converted into a sequence of images that are stored as joint photographic experts group (jpeg) files. The background is subtracted from each image using a …


Sources Of Variability In A Proteomic Experiment, Scott Daniel Crawford Aug 2006

Sources Of Variability In A Proteomic Experiment, Scott Daniel Crawford

Theses and Dissertations

The study of proteomics holds the hope for detecting serious diseases earlier than is currently possible by analyzing blood samples in a mass spectrometer. Unfortunately, the statistics involved in comparing a control group to a diseased group are not trivial, and these difficulties have led others to incorrect decisions in the past. This paper considers a nested design that was used to quantify and identify the sources of variation in the mass spectrometer at BYU, so that correct conclusions can be drawn from blood samples analyzed in proteomics. Algorithms were developed which detect, align, correct, and cluster the peaks in …


Understanding Brigham Young University's Technology Teacher Education Program's Sucess In Attracting And Retaining Female Students, Katrina M. Cox Jul 2006

Understanding Brigham Young University's Technology Teacher Education Program's Sucess In Attracting And Retaining Female Students, Katrina M. Cox

Theses and Dissertations

The purpose of the study was to attempt to understand why Brigham Young University Technology Teacher Education program has attracted and retained a high number of females. This was done through a self-created survey composed of four forced responses, distributed among the Winter 2006 semester students. Likert-scale questions were outlined according to the five theoretical influences on women in technology, as established by Welty and Puck (2001) and two of the three relationships of academia, as established by Haynie III (1999), as well as three free response questions regarding retention and attraction within the major. Findings suggested strong positive polarity …


Computation Of Weights For Probabilistic Record Linkage Using The Em Algorithm, G. John Bauman Jun 2006

Computation Of Weights For Probabilistic Record Linkage Using The Em Algorithm, G. John Bauman

Theses and Dissertations

Record linkage is the process of combining information about a single individual from two or more records. Probabilistic record linkage gives weights to each field that is compared. The decision of whether the records should be linked is then determined by the sum of the weights, or “Score”, over all fields compared. Using methods similar to the simple versus simple most powerful test, an optimal record linkage decision rule can be established to minimize the number of unlinked records when the probability of false positive and false negative errors are specified. The weights needed for probabilistic record linkage necessitate linking …


Small Sample Confidence Intervals In Log Space Back-Transformed From Normal Space, Jason E. Tisdel Jun 2006

Small Sample Confidence Intervals In Log Space Back-Transformed From Normal Space, Jason E. Tisdel

Theses and Dissertations

The logarithmic transformation is commonly applied to a lognormal data set to improve symmetry, homoscedasticity, and linearity. Simple to implement and easy to understand, the logarithm function transforms the original data to closely resemble a normal distribution. Analysis in the normal space provides point estimates and confidence intervals, but transformation back to the original space using the naive approach yields confidence intervals of impractical width. The naive approach offers results that are often inadequate for practical purpose. We present an alternative approach that provides improved results in the form of decreased interval width, increased confidence level, or both. Our alternative …


Food Shelf Life: Estimation And Experimental Design, Ross Allen Andrew Larsen May 2006

Food Shelf Life: Estimation And Experimental Design, Ross Allen Andrew Larsen

Theses and Dissertations

Shelf life is a parameter of the lifetime distribution of a food product, usually the time until a specified proportion (1-50%) of the product has spoiled according to taste. The data used to estimate shelf life typically come from a planned experiment with sampled food items observed at specified times. The observation times are usually selected adaptively using ‘staggered sampling.’ Ad-hoc methods based on linear regression have been recommended to estimate shelf life. However, other methods based on maximizing a likelihood (MLE) have been proposed, studied, and used. Both methods assume the Weibull distribution. The observed lifetimes in shelf life …


Bayesian And Positive Matrix Factorization Approaches To Pollution Source Apportionment, Jeff William Lingwall May 2006

Bayesian And Positive Matrix Factorization Approaches To Pollution Source Apportionment, Jeff William Lingwall

Theses and Dissertations

The use of Positive Matrix Factorization (PMF) in pollution source apportionment (PSA) is examined and illustrated. A study of its settings is conducted in order to optimize them in the context of PSA. The use of a priori information in PMF is examined, in the form of target factor profiles and pulling profile elements to zero. A Bayesian model using lognormal prior distributions for source profiles and source contributions is fit and examined.


The Longitudinal Effect Of Self-Monitoring And Locus Of Control On Social Network Position In Friendship Networks, Gary J. Moore Mar 2006

The Longitudinal Effect Of Self-Monitoring And Locus Of Control On Social Network Position In Friendship Networks, Gary J. Moore

Theses and Dissertations

The purpose of this research was to identify how enduring personality characteristics predict a person's location in a network, locations which in turn affect outcomes such as performance. Specifically, this thesis examines how self-monitoring and locus of control influence an individual's location in a friendship social network over time. Hierarchical Linear Modeling (HLM) was used to analyze 28 groups of students and instructors at a military training course over six and one half weeks. Self-monitoring predicted betweenness centrality in five of six time periods while locus of control predicted betweenness centrality in three of six time periods. The moderation of …


Modeling The Performance Of A Baseball Player's Offensive Production, Michael Ross Smith Mar 2006

Modeling The Performance Of A Baseball Player's Offensive Production, Michael Ross Smith

Theses and Dissertations

This project addresses the problem of comparing the offensive abilities of players from different eras in Major League Baseball (MLB). We will study players from the perspective of an overall offensive summary statistic that is highly linked with scoring runs, or the Berry Value. We will build an additive model to estimate the innate ability of the player, the effect of the relative level of competition of each season, and the effect of age on performance using piecewise age curves. Using Hierarchical Bayes methodology with Gibbs sampling, we model each of these effects for each individual. The results of the …


Optimization Of A Multi-Echelon Repair System Via Generalized Pattern Search With Ranking And Selection: A Computational Study, Derek D. Tharaldson Mar 2006

Optimization Of A Multi-Echelon Repair System Via Generalized Pattern Search With Ranking And Selection: A Computational Study, Derek D. Tharaldson

Theses and Dissertations

With increasing developments in computer technology and available software, simulation is becoming a widely used tool to model, analyze, and improve a real world system or process. However, simulation in itself is not an optimization approach. Common optimization procedures require either an explicit mathematical formulation or numerous function evaluations at improving iterative points. Mathematical formulation is generally impossible for problems where simulation is relevant, which are characteristically the types of problems that arise in practical applications. Further complicating matters is the variability in the simulation response which can cause problems in iterative techniques using the simulation model as a function …


Correlating Factors Between Student Participation And Student Learning Via A Service Learning Project In Secondary Education: A Case Study, Shawn V. Jensen Jan 2006

Correlating Factors Between Student Participation And Student Learning Via A Service Learning Project In Secondary Education: A Case Study, Shawn V. Jensen

Theses and Dissertations

In this study a service-learning project was conducted with secondary students in a construction technology based course. Three research questions were considered; (1) does service learning projects help to engage student participation, (2) can students learn the course curriculum while participating in a service learning project, and (3) is there a correlation between student participation and student learning as it pertains to service learning projects? The data was collected through surveys, observations, interviews, and evaluations. The researcher concluded the following from the study; 92% of the students were actively participating in the two week service project, 76% of the students …


A Normal-Mixture Model With Random-Effects For Rr-Interval Data, Jessica Mckinney Ketchum Jan 2006

A Normal-Mixture Model With Random-Effects For Rr-Interval Data, Jessica Mckinney Ketchum

Theses and Dissertations

In many applications of random-effects models to longitudinal data, such as heart rate variability (HRV) data, a normal-mixture distribution seems to be more appropriate than the normal distribution assumption. While the random-effects methodology is well developed for several distributions in the exponential family, the case of the normal-mixture has not been dealt with adequately in the literature. The models and the estimation methods that have been proposed in the past assume the conditional model (fixing the random-effects) to be normal and allow a mixture distribution for the random effects (Xu and Hedeker, 2001, Xu, 1995). The methods proposed in this …


Meta-Analysis Of Open Vs Closed Surgery Of Mandibular Condyle Fractures, Marcy Lauren Nussbaum Jan 2006

Meta-Analysis Of Open Vs Closed Surgery Of Mandibular Condyle Fractures, Marcy Lauren Nussbaum

Theses and Dissertations

A review of the literature reveals a difference of opinion regarding whether open or closed reduction of condylar fractures produces the best results. It would be beneficial to critically analyze past studies that have directly compared the two methods in an attempt to answer this question. A Medline search for articles using the key words 'mandibular condyle fractures' and 'mandibular condyle fractures surgery' was performed. The articles chosen for the meta-analysis contained data on at least one of the following: postoperative maximum mouth opening, lateral excursion, protrusion, deviation on opening, asymmetry, and joint pain or muscle pain. Several common statistical …


Optimal Clustering: Genetic Constrained K-Means And Linear Programming Algorithms, Jianmin Zhao Jan 2006

Optimal Clustering: Genetic Constrained K-Means And Linear Programming Algorithms, Jianmin Zhao

Theses and Dissertations

Methods for determining clusters of data under- specified constraints have recently gained popularity. Although general constraints may be used, we focus on clustering methods with the constraint of a minimal cluster size. In this dissertation, we propose two constrained k-means algorithms: Linear Programming Algorithm (LPA) and Genetic Constrained K-means Algorithm (GCKA). Linear Programming Algorithm modifies the k-means algorithm into a linear programming problem with constraints requiring that each cluster have m or more subjects. In order to achieve an acceptable clustering solution, we run the algorithm with a large number of random sets of initial seeds, and choose the solution …


Assessing, Modifying, And Combining Data Fields From The Virginia Office Of The Chief Medical Examiner (Ocme) Dataset And The Virginia Department Of Forensic Science (Dfs) Datasets In Order To Compare Concentrations Of Selected Drugs, Amy Elizabeth Herrin Jan 2006

Assessing, Modifying, And Combining Data Fields From The Virginia Office Of The Chief Medical Examiner (Ocme) Dataset And The Virginia Department Of Forensic Science (Dfs) Datasets In Order To Compare Concentrations Of Selected Drugs, Amy Elizabeth Herrin

Theses and Dissertations

The Medical Examiner of Virginia (ME) dataset and the Virginia Department of Forensic Science Driving Under the Influence of Drugs (DUI) datasets were used to determine whether people have the potential to develop tolerances to diphenhydramine, cocaine, oxycodone, hydrocodone, methadone, and morphine. These datasets included the years 2000-2004 and were used to compare the concentrations of these six drugs between people who died from a drug-related cause of death (of the drug of interest) and people who were pulled over for driving under the influence. Three drug pattern groups were created to divide each of the six drug-specific datasets in …


Design And Analysis Methods For Cluster Randomized Trials With Pair-Matching On Baseline Outcome: Reduction Of Treatment Effect Variance, Misook Park Jan 2006

Design And Analysis Methods For Cluster Randomized Trials With Pair-Matching On Baseline Outcome: Reduction Of Treatment Effect Variance, Misook Park

Theses and Dissertations

Cluster randomized trials (CRT) are comparative studies designed to evaluate interventions where the unit of analysis and randomization is the cluster but the unit of observation is individuals within clusters. Typically such designs involve a limited number of clusters and thus the variation between clusters is left uncontrolled. Experimental designs and analysis strategies that minimize this variance are required. In this work we focus on the CRT with pre-post intervention measures. By incorporating the baseline measure into the analysis, we can effectively reduce the variance of the treatment effect. Well known methods such as adjustment for baseline as a covariate …


A Comparison For Longitudinal Data Missing Due To Truncation, Rong Liu Jan 2006

A Comparison For Longitudinal Data Missing Due To Truncation, Rong Liu

Theses and Dissertations

Many longitudinal clinical studies suffer from patient dropout. Often the dropout is nonignorable and the missing mechanism needs to be incorporated in the analysis. The methods handling missing data make various assumptions about the missing mechanism, and their utility in practice depends on whether these assumptions apply in a specific application. Ramakrishnan and Wang (2005) proposed a method (MDT) to handle nonignorable missing data, where missing is due to the observations exceeding an unobserved threshold. Assuming that the observations arise from a truncated normal distribution, they suggested an EM algorithm to simplify the estimation.In this dissertation the EM algorithm is …


Statistical Methods And Experimental Design For Inference Regarding Dose And/Or Interaction Thresholds Along A Fixed-Ratio Ray, Sharon Dziuba Yeatts Jan 2006

Statistical Methods And Experimental Design For Inference Regarding Dose And/Or Interaction Thresholds Along A Fixed-Ratio Ray, Sharon Dziuba Yeatts

Theses and Dissertations

An alternative to the full factorial design, the ray design is appropriate for investigating a mixture of c chemicals, which are present according to a fixed mixing ratio, called the mixture ray. Using single chemical and mixture ray data, we can investigate interaction among the chemicals in a particular mixture. Statistical models have been used to describe the dose-response relationship of the single agents and the mixture; additivity is tested through the significance of model parameters associated with the coincidence of the additivity and mixture models.It is often assumed that a chemical or mixture must be administered above an unknown …


Quantifying The Effects Of Correlated Covariates On Variable Importance Estimates From Random Forests, Ryan Vincent Kimes Jan 2006

Quantifying The Effects Of Correlated Covariates On Variable Importance Estimates From Random Forests, Ryan Vincent Kimes

Theses and Dissertations

Recent advances in computing technology have lead to the development of algorithmic modeling techniques. These methods can be used to analyze data which are difficult to analyze using traditional statistical models. This study examined the effectiveness of variable importance estimates from the random forest algorithm in identifying the true predictor among a large number of candidate predictors. A simulation study was conducted using twenty different levels of association among the independent variables and seven different levels of association between the true predictor and the response. We conclude that the random forest method is an effective classification tool when the goals …