Open Access. Powered by Scholars. Published by Universities.®

Physical Sciences and Mathematics Commons

Open Access. Powered by Scholars. Published by Universities.®

Articles 1 - 26 of 26

Full-Text Articles in Physical Sciences and Mathematics

Innate Immunity, The Hepatic Extracellular Matrix, And Liver Injury: Mathematical Modeling Of Metastatic Potential And Tumor Development In Alcoholic Liver Disease., Shanice V. Hudson Dec 2018

Innate Immunity, The Hepatic Extracellular Matrix, And Liver Injury: Mathematical Modeling Of Metastatic Potential And Tumor Development In Alcoholic Liver Disease., Shanice V. Hudson

Electronic Theses and Dissertations

The overarching goals of the current work are to fill key gaps in the current understanding of alcohol consumption and the risk of metastasis to the liver. Considering the evidence this research group has compiled confirming that the hepatic matrisome responds dynamically to injury, an altered extracellular matrix (ECM) profile appears to be a key feature of pre-fibrotic inflammatory injury in the liver. This group has demonstrated that the hepatic ECM responds dynamically to alcohol exposure, in particular, sensitizing the liver to LPS-induced inflammatory damage. Although the study of alcohol in its role as a contributing factor to oncogenesis and …


Effectiveness Of Prescribed Fire On Meeting Fuel Load And Wildlife Habitat Management Objectives In East Texas National Forests, Trey Wall Dec 2018

Effectiveness Of Prescribed Fire On Meeting Fuel Load And Wildlife Habitat Management Objectives In East Texas National Forests, Trey Wall

Electronic Theses and Dissertations

Using standardized methodology outlined by the United States Forest Service and the National Forests and Grasslands in Texas’ Fire Monitoring Program for data collection, the efficacy of current Forest Service prescribed burn regimes were analyzed for 24 study sites in East Texas National Forests. Study sites were located within Sam Houston, Davy Crockett, and Angelina/Sabine National Forests. Efficacy was determined by comparing defined management objectives established by the Forest Service to the data collected at the study sites. The results conclude that most objectives, as outlined by the Forest Service, are not being met with the current practices. Re-visitation of …


Wald Confidence Intervals For A Single Poisson Parameter And Binomial Misclassification Parameter When The Data Is Subject To Misclassification, Nishantha Janith Chandrasena Poddiwala Hewage Aug 2018

Wald Confidence Intervals For A Single Poisson Parameter And Binomial Misclassification Parameter When The Data Is Subject To Misclassification, Nishantha Janith Chandrasena Poddiwala Hewage

Electronic Theses and Dissertations

This thesis is based on a Poisson model that uses both error-free data and error-prone data subject to misclassification in the form of false-negative and false-positive counts. We present maximum likelihood estimators (MLEs), Fisher's Information, and Wald statistics for Poisson rate parameter and the two misclassification parameters. Next, we invert the Wald statistics to get asymptotic confidence intervals for Poisson rate parameter and false-negative rate parameter. The coverage and width properties for various sample size and parameter configurations are studied via a simulation study. Finally, we apply the MLEs and confidence intervals to one real data set and another realistic …


Clustering Mixed Data: An Extension Of The Gower Coefficient With Weighted L2 Distance, Augustine Oppong Aug 2018

Clustering Mixed Data: An Extension Of The Gower Coefficient With Weighted L2 Distance, Augustine Oppong

Electronic Theses and Dissertations

Sorting out data into partitions is increasing becoming complex as the constituents of data is growing outward everyday. Mixed data comprises continuous, categorical, directional functional and other types of variables. Clustering mixed data is based on special dissimilarities of the variables. Some data types may influence the clustering solution. Assigning appropriate weight to the functional data may improve the performance of the clustering algorithm. In this paper we use the extension of the Gower coefficient with judciously chosen weight for the L2 to cluster mixed data.The benefits of weighting are demonstrated both in in applications to the Buoy data set …


The Expected Number Of Patterns In A Random Generated Permutation On [N] = {1,2,...,N}, Evelyn Fokuoh Aug 2018

The Expected Number Of Patterns In A Random Generated Permutation On [N] = {1,2,...,N}, Evelyn Fokuoh

Electronic Theses and Dissertations

Previous work by Flaxman (2004) and Biers-Ariel et al. (2018) focused on the number of distinct words embedded in a string of words of length n. In this thesis, we will extend this work to permutations, focusing on the maximum number of distinct permutations contained in a permutation on [n] = {1,2,...,n} and on the expected number of distinct permutations contained in a random permutation on [n]. We further considered the problem where repetition of subsequences are as a result of the occurrence of (Type A and/or Type B) replications. Our method of enumerating the Type A replications causes double …


Distribution Of A Sum Of Random Variables When The Sample Size Is A Poisson Distribution, Mark Pfister Aug 2018

Distribution Of A Sum Of Random Variables When The Sample Size Is A Poisson Distribution, Mark Pfister

Electronic Theses and Dissertations

A probability distribution is a statistical function that describes the probability of possible outcomes in an experiment or occurrence. There are many different probability distributions that give the probability of an event happening, given some sample size n. An important question in statistics is to determine the distribution of the sum of independent random variables when the sample size n is fixed. For example, it is known that the sum of n independent Bernoulli random variables with success probability p is a Binomial distribution with parameters n and p: However, this is not true when the sample size …


Bayesian Analytical Approaches For Metabolomics : A Novel Method For Molecular Structure-Informed Metabolite Interaction Modeling, A Novel Diagnostic Model For Differentiating Myocardial Infarction Type, And Approaches For Compound Identification Given Mass Spectrometry Data., Patrick J. Trainor Aug 2018

Bayesian Analytical Approaches For Metabolomics : A Novel Method For Molecular Structure-Informed Metabolite Interaction Modeling, A Novel Diagnostic Model For Differentiating Myocardial Infarction Type, And Approaches For Compound Identification Given Mass Spectrometry Data., Patrick J. Trainor

Electronic Theses and Dissertations

Metabolomics, the study of small molecules in biological systems, has enjoyed great success in enabling researchers to examine disease-associated metabolic dysregulation and has been utilized for the discovery biomarkers of disease and phenotypic states. In spite of recent technological advances in the analytical platforms utilized in metabolomics and the proliferation of tools for the analysis of metabolomics data, significant challenges in metabolomics data analyses remain. In this dissertation, we present three of these challenges and Bayesian methodological solutions for each. In the first part we develop a new methodology to serve a basis for making higher order inferences in metabolomics, …


Generalized Spatiotemporal Modeling And Causal Inference For Assessing Treatment Effects For Multiple Groups For Ordinal Outcome., Soutik Ghosal Aug 2018

Generalized Spatiotemporal Modeling And Causal Inference For Assessing Treatment Effects For Multiple Groups For Ordinal Outcome., Soutik Ghosal

Electronic Theses and Dissertations

This dissertation consists of three projects and can be categorized in two broad research areas: generalized spatiotemporal modeling and causal inference based on observational data. In the first project, I introduce a Bayesian hierarchical mixed effect hurdle model with a nested random effect structure to model the count for primary care providers and understand their spatial and temporal variation. This study further enables us to identify the health professional shortage areas and the possible impacting factors. In the second project, I have unified popular parametric and nonparametric propensity score-based methods to assess the treatment effect of multiple groups for ordinal …


Spatio-Temporal Dynamics Of Atlantic Cod Bycatch In The Maine Lobster Fishery And Its Impacts On Stock Assessment, Robert E. Boenish May 2018

Spatio-Temporal Dynamics Of Atlantic Cod Bycatch In The Maine Lobster Fishery And Its Impacts On Stock Assessment, Robert E. Boenish

Electronic Theses and Dissertations

Of the most iconic fish species in the world, the Atlantic cod (Gadus morhua, hereafter, cod) has been a mainstay in the North Atlantic for centuries. While many global fish stocks have received increased pressure with the advent of new, more efficient fishing technology in the mid-20th century, exceptional pressure has been placed on this prized gadoid. Bycatch, or the unintended catch of organisms, is one of the biggest global fisheries issues. Directly resulting from the failed recovery of cod in the GoM, attention has been placed as to possible sources of unaccounted catch. Among the most …


Re-Evaluating Performance Measurement: New Mathematical Methods To Address Common Performance Measurement Challenges, Jordan David Benis May 2018

Re-Evaluating Performance Measurement: New Mathematical Methods To Address Common Performance Measurement Challenges, Jordan David Benis

Electronic Theses and Dissertations

Performance Measurement is an essential discipline for any business. Robust and reliable performance metrics for people, processes, and technologies enable a business to identify and address deficiencies to improve performance and profitability. The complexity of modern operating environments presents real challenges to developing equitable and accurate performance metrics. This thesis explores and develops two new methods to address common challenges encountered in businesses across the world. The first method addresses the challenge of estimating the relative complexity of various tasks by utilizing the Pearson Correlation Coefficient to identify potentially over weighted and under weighted tasks. The second method addresses the …


Designing A Calibration Set In Spectral Space For Efficient Development Of An Nir Method For Tablet Analysis, Md Anik Alam May 2018

Designing A Calibration Set In Spectral Space For Efficient Development Of An Nir Method For Tablet Analysis, Md Anik Alam

Electronic Theses and Dissertations

Designing a calibration set is the first step in developing a spectroscopic calibration method for quantitative analysis of pharmaceutical tablets. This step is critical because successful model development depends on the suitability of the calibration data. For spectroscopic-based methods, traditional concentration based techniques for designing calibration sets are prone to have redundant information while simultaneously lacking necessary information for a successful calibration model. The traditional method also follows the same design approach for different spectroscopic techniques and different formulations, thereby lacks the optimizing capability to be technique and formulation specific.

A method for designing a calibration set in the Near …


Geostatistical Analysis Of Potential Sinkhole Risk: Examining Spatial And Temporal Climate Relationships In Tennessee And Florida, Kimberly Blazzard May 2018

Geostatistical Analysis Of Potential Sinkhole Risk: Examining Spatial And Temporal Climate Relationships In Tennessee And Florida, Kimberly Blazzard

Electronic Theses and Dissertations

Sinkholes are a significant hazard for the southeastern United States. Although differences in climate are known to affect karst environments differently, quantitative analyses correlating sinkhole formation with climate variables is lacking. A temporal linear regression for Florida sinkholes and two modeled regressions for Tennessee sinkholes were produced: a general linearized logistic regression and a MaxEnt derived species distribution model. Temporal results showed highly significant correlations with precipitation, teleconnection patterns, temperature, and CO2, while spatial results showed highly significant correlations with precipitation, wind speed, solar radiation, and maximum temperature. Regression results indicated that some sinkhole formation variability could be …


Multi Self-Adapting Particle Swarm Optimization Algorithm (Msapso)., Gerhard Koch May 2018

Multi Self-Adapting Particle Swarm Optimization Algorithm (Msapso)., Gerhard Koch

Electronic Theses and Dissertations

The performance and stability of the Particle Swarm Optimization algorithm depends on parameters that are typically tuned manually or adapted based on knowledge from empirical parameter studies. Such parameter selection is ineffectual when faced with a broad range of problem types, which often hinders the adoption of PSO to real world problems. This dissertation develops a dynamic self-optimization approach for the respective parameters (inertia weight, social and cognition). The effects of self-adaption for the optimal balance between superior performance (convergence) and the robustness (divergence) of the algorithm with regard to both simple and complex benchmark functions is investigated. This work …


Evaluation Of Using The Bootstrap Procedure To Estimate The Population Variance, Nghia Trong Nguyen May 2018

Evaluation Of Using The Bootstrap Procedure To Estimate The Population Variance, Nghia Trong Nguyen

Electronic Theses and Dissertations

The bootstrap procedure is widely used in nonparametric statistics to generate an empirical sampling distribution from a given sample data set for a statistic of interest. Generally, the results are good for location parameters such as population mean, median, and even for estimating a population correlation. However, the results for a population variance, which is a spread parameter, are not as good due to the resampling nature of the bootstrap method. Bootstrap samples are constructed using sampling with replacement; consequently, groups of observations with zero variance manifest in these samples. As a result, a bootstrap variance estimator will carry a …


Longitudinal Tracking Of Physiological State With Electromyographic Signals., Robert Warren Stallard May 2018

Longitudinal Tracking Of Physiological State With Electromyographic Signals., Robert Warren Stallard

Electronic Theses and Dissertations

Electrophysiological measurements have been used in recent history to classify instantaneous physiological configurations, e.g., hand gestures. This work investigates the feasibility of working with changes in physiological configurations over time (i.e., longitudinally) using a variety of algorithms from the machine learning domain. We demonstrate a high degree of classification accuracy for a binary classification problem derived from electromyography measurements before and after a 35-day bedrest. The problem difficulty is increased with a more dynamic experiment testing for changes in astronaut sensorimotor performance by taking electromyography and force plate measurements before, during, and after a jump from a small platform. A …


Improving The Detection Limit Of Tau Aggregates For Use With Biological Samples, Emily Rickman Hager Jan 2018

Improving The Detection Limit Of Tau Aggregates For Use With Biological Samples, Emily Rickman Hager

Electronic Theses and Dissertations

The protein Tau is found in neurofibrillary tangles in Alzheimer's disease and over 20 other neurodegenerative diseases. An assay has been developed to detect minute amounts of fibrils from human brain tissue. This assay subjects brain tissue extract and recombinant Tau to several rounds of sonication and incubation. Incubation allows recombinant Tau to add itself to the ends of the existing fibrils in brain tissue extract. Sonication breaks the existing fibrils in the brain tissue extract offering more ends for Tau to add onto. Cycles of sonication and incubation have been shown to allow for amplification of Tau fibrils from …


Location Optimization Of A Coal Power Plant To Balance Coal Supply And Electric Transmission Costs Against Plant’S Emission Exposure, Najam Khan Jan 2018

Location Optimization Of A Coal Power Plant To Balance Coal Supply And Electric Transmission Costs Against Plant’S Emission Exposure, Najam Khan

Electronic Theses and Dissertations

This research is focused on developing a location analysis methodology that can minimize the pollutant exposure to the public while ensuring that the combined costs of electric transmission losses and coal logistics are minimized. Coal power plants will provide a critical contribution towards meeting electricity demands for various nations in the foreseeable future. The site selection for a new coal power plant is extremely important from an investment point of view. The operational costs for running a coal power plant can be minimized by a combined emphasis on placing a coal power plant near coal mines as well as customers. …


Statistical Algorithms And Bioinformatics Tools Development For Computational Analysis Of High-Throughput Transcriptomic Data, Adam Mcdermaid Jan 2018

Statistical Algorithms And Bioinformatics Tools Development For Computational Analysis Of High-Throughput Transcriptomic Data, Adam Mcdermaid

Electronic Theses and Dissertations

Next-Generation Sequencing technologies allow for a substantial increase in the amount of data available for various biological studies. In order to effectively and efficiently analyze this data, computational approaches combining mathematics, statistics, computer science, and biology are implemented. Even with the substantial efforts devoted to development of these approaches, numerous issues and pitfalls remain. One of these issues is mapping uncertainty, in which read alignment results are biased due to the inherent difficulties associated with accurately aligning RNA-Sequencing reads. GeneQC is an alignment quality control tool that provides insight into the severity of mapping uncertainty in each annotated gene from …


Variable Selection Techniques For Clustering On The Unit Hypersphere, Damon Bayer Jan 2018

Variable Selection Techniques For Clustering On The Unit Hypersphere, Damon Bayer

Electronic Theses and Dissertations

Mixtures of von Mises-Fisher distributions have been shown to be an effective model for clustering data on a unit hypersphere, but variable selection for these models remains an important and challenging problem. In this paper, we derive two variants of the expectation-maximization framework, which are each used to identify a specific type of irrelevant variables for these models. The first type are noise variables, which are not useful for separating any pairs of clusters. The second type are redundant variables, which may be useful for separating pairs of clusters, but do not enable any additional separation beyond the separability provided …


The Impact Of Data Sovereignty On American Indian Self-Determination: A Framework Proof Of Concept Using Data Science, Joseph Carver Robertson Jan 2018

The Impact Of Data Sovereignty On American Indian Self-Determination: A Framework Proof Of Concept Using Data Science, Joseph Carver Robertson

Electronic Theses and Dissertations

The Data Sovereignty Initiative is a collection of ideas that was designed to create SMART solutions for tribal communities. This concept was to develop a horizontal governance framework to create a strategic act of sovereignty using data science. The core concept of this idea was to present data sovereignty as a way for tribal communities to take ownership of data in order to affect policy and strategic decisions that are data driven in nature. The case studies in this manuscript were developed around statistical theories of spatial statistics, exploratory data analysis, and machine learning. And although these case studies are …


Development Of Biclustering Techniques For Gene Expression Data Modeling And Mining, Juan Xie Jan 2018

Development Of Biclustering Techniques For Gene Expression Data Modeling And Mining, Juan Xie

Electronic Theses and Dissertations

The next-generation sequencing technologies can generate large-scale biological data with higher resolution, better accuracy, and lower technical variation than the arraybased counterparts. RNA sequencing (RNA-Seq) can generate genome-scale gene expression data in biological samples at a given moment, facilitating a better understanding of cell functions at genetic and cellular levels. The abundance of gene expression datasets provides an opportunity to identify genes with similar expression patterns across multiple conditions, i.e., co-expression gene modules (CEMs). Genomescale identification of CEMs can be modeled and solved by biclustering, a twodimensional data mining technique that allows clustering of rows and columns in a gene …


Old English Character Recognition Using Neural Networks, Sattajit Sutradhar Jan 2018

Old English Character Recognition Using Neural Networks, Sattajit Sutradhar

Electronic Theses and Dissertations

Character recognition has been capturing the interest of researchers since the beginning of the twentieth century. While the Optical Character Recognition for printed material is very robust and widespread nowadays, the recognition of handwritten materials lags behind. In our digital era more and more historical, handwritten documents are digitized and made available to the general public. However, these digital copies of handwritten materials lack the automatic content recognition feature of their printed materials counterparts. We are proposing a practical, accurate, and computationally efficient method for Old English character recognition from manuscript images. Our method relies on a modern machine learning …


Examining The Issue Of Compliance With Personal Protective Equipment Among Wastewater Workers Across The Southeast Region Of The United States, Tamara L. Wright Jan 2018

Examining The Issue Of Compliance With Personal Protective Equipment Among Wastewater Workers Across The Southeast Region Of The United States, Tamara L. Wright

Electronic Theses and Dissertations

Wastewater workers are exposed to different occupational hazards such as chemicals, gases, viruses, and bacteria. Personal Protective Equipment (PPE) is a significant factor that can reduce or increase the probability of an accident from hazardous exposures to chemicals and microbial contaminants. The purpose of this study was to identify wastewater worker’s beliefs and practices on wearing PPE and protections offered by PPE through the integration of the Health Belief Model (HBM). Participants were workers in the wastewater industry, which included wastewater operators, laboratory analysts, maintenance workers, wastewater collection workers, equipment operators, managers, and supervisors (n=272). The instrument was a self-administered …


Some New And Generalized Distributions Via Exponentiation, Gamma And Marshall-Olkin Generators With Applications, Hameed Abiodun Jimoh Jan 2018

Some New And Generalized Distributions Via Exponentiation, Gamma And Marshall-Olkin Generators With Applications, Hameed Abiodun Jimoh

Electronic Theses and Dissertations

Three new generalized distributions developed via completing risk, gamma generator, Marshall-Olkin generator and exponentiation techniques are proposed and studied. Structural properties including quantile functions, hazard rate functions, moment, conditional moments, mean deviations, R\'enyi entropy, distribution of order statistics and maximum likelihood estimates are presented. Monte Carlo simulation is employed to examine the performance of the proposed distributions. Applications of the generalized distributions to real lifetime data are presented to illustrate the usefulness of the models.


A Comparison Of Bridge Deterioration Models, Toktam Naderimoghaddam Jan 2018

A Comparison Of Bridge Deterioration Models, Toktam Naderimoghaddam

Electronic Theses and Dissertations

Predicting how bridges will deteriorate is the key to budgeting financial and personnel resources. Deterioration models exist for specific components of a bridge, but no models exist for the sufficiency rating which is an overall measure of the condition and relevance of a bridge used for determining eligibility for federal funds.

We have 25 years worth of data collected by the Georgia Department of Transportation from 1992 to 2016 about all bridges in the State of Georgia. More precisely, each row in this data set includes the characteristics of each bridge along with the sufficiency rating of that bridge in …


Multiclass Classification Using Support Vector Machines, Duleep Prasanna W. Rathgamage Don Jan 2018

Multiclass Classification Using Support Vector Machines, Duleep Prasanna W. Rathgamage Don

Electronic Theses and Dissertations

In this thesis, we discuss different SVM methods for multiclass classification and introduce the Divide and Conquer Support Vector Machine (DCSVM) algorithm which relies on data sparsity in high dimensional space and performs a smart partitioning of the whole training data set into disjoint subsets that are easily separable. A single prediction performed between two partitions eliminates one or more classes in a single partition, leaving only a reduced number of candidate classes for subsequent steps. The algorithm continues recursively, reducing the number of classes at each step until a final binary decision is made between the last two classes …