Open Access. Powered by Scholars. Published by Universities.®

Digital Commons Network

Open Access. Powered by Scholars. Published by Universities.®

Articles 1 - 20 of 20

Full-Text Articles in Entire DC Network

Characterizations And Reliability Measures Of The Generalized Log Burr Xii Distribution, Fiaz Ahmad Bhatti, Gholamhossein G. Hamedani, Azeem Ali, Sedigheh Mirzaei Salehabadi, Munir Ahmad Jul 2021

Characterizations And Reliability Measures Of The Generalized Log Burr Xii Distribution, Fiaz Ahmad Bhatti, Gholamhossein G. Hamedani, Azeem Ali, Sedigheh Mirzaei Salehabadi, Munir Ahmad

Mathematical and Statistical Science Faculty Research and Publications

In this paper, we derive the generalized log Burr XII (GLBXII) distribution [2] from the generalized Burr-Hatke differential equation. We characterize the GLBXII distribution via innovative techniques. We derive various reliability measures (series and parallel). We also authenticate the potentiality of the GLBXII model via economics applications. The applications of characterizations and reliability measures of the GLBXII distribution in different disciplines of science will be profitable for scientists.


Estimating The Reliability Of A Component Between Two Stresses From Gompertz-Frechet Model, Sarah Adnan Jabr, Nada Sabah Karam Apr 2021

Estimating The Reliability Of A Component Between Two Stresses From Gompertz-Frechet Model, Sarah Adnan Jabr, Nada Sabah Karam

Al-Qadisiyah Journal of Pure Science

In this paper, the reliability of the stress-strength model is derived for probability P(


A Novel Approach For Assessing The Reliability Of Data Contained In A Single Valued Neutrosophic Number And Its Application In Multiple Criteria Decision Making, Florentin Smarandache, Dragisa Stanujkic, Darjan Karabasevic, Gabrijela Popovic Sep 2020

A Novel Approach For Assessing The Reliability Of Data Contained In A Single Valued Neutrosophic Number And Its Application In Multiple Criteria Decision Making, Florentin Smarandache, Dragisa Stanujkic, Darjan Karabasevic, Gabrijela Popovic

Branch Mathematics and Statistics Faculty and Staff Publications

Multiple criteria decision making is one of the many areas where neutrosophic sets have been applied to solve various problems so far.


New Modified Singh-Maddala Distribution: Development, Properties, Characterizations And Applications, Fiaz Ahmad Bhatti, Gholamhossein G. Hamedani, Mustafa Ç. Korkmaz, Munir Ahmad Jul 2019

New Modified Singh-Maddala Distribution: Development, Properties, Characterizations And Applications, Fiaz Ahmad Bhatti, Gholamhossein G. Hamedani, Mustafa Ç. Korkmaz, Munir Ahmad

Mathematical and Statistical Science Faculty Research and Publications

In this paper, a new five-parameter extended Burr XII model called new modified Singh-Maddala (NMSM) is developed from cumulative hazard function of the modified log extended integrated beta hazard (MLEIBH) model. The NMSM density function is left-skewed, right-skewed and symmetrical. The Lambert W function is used to study descriptive measures based on quantile, moments, and moments of order statistics, incomplete moments, inequality measures and residual life function. Different reliability and uncertainty measures are also theoretically established. The NMSM distribution is characterized via different techniques and its parameters are estimated using maximum likelihood method. The simulation studies are performed on the …


Pm2.5 Data Reliability And Air Quality Improvement Trends In Beijing, Huimin Li May 2019

Pm2.5 Data Reliability And Air Quality Improvement Trends In Beijing, Huimin Li

Theses and Dissertations

PM2.5 has been a main environmental concern due to its adverse effects on human health and society. We used data from two sources: monitoring station of the U.S. Embassy in Beijing, and several nearby monitoring stations of the Chinese Ministry of Environmental Protection. This study includes investigating (1) PM2.5 historical data reliability, (2) PM2.5 real-time data reliability, and (3) air quality improvement trends in Beijing over the past decade. We used graphical methods, descriptive statistics, correlation analysis, and inferential analyses including paired samples t-test, ANOVA, and Kruskal-Wallis test. We reported effect sizes to aid study on practical significance. Inferential procedures' …


Cronbach’S Alpha Under Insufficient Effort Responding: An Analytic Approach, Stephen W. Carden, Trevor R. Camper, Nicholas S. Holtzman Jan 2019

Cronbach’S Alpha Under Insufficient Effort Responding: An Analytic Approach, Stephen W. Carden, Trevor R. Camper, Nicholas S. Holtzman

Department of Psychology Faculty Publications

Surveys commonly suffer from insufficient effort responding (IER). If not accounted for, IER can cause biases and lead to false conclusions. In particular, Cronbach’s alpha has been empirically observed to either deflate or inflate due to IER. This paper will elucidate how IER impacts Cronbach’s alpha in a variety of situations. Previous results concerning internal consistency under mixture models are extended to obtain a characterization of Cronbach’s alpha in terms of item validities, average variances, and average covariances. The characterization is then applied to contaminating distributions representing various types of IER. The discussion will provide commentary on previous simulation-based investigations, …


Quantifying Similarity In Reliability Surfaces Using The Probability Of Agreement, Nathaniel Stevens, C. M. Anderson-Cook Jan 2017

Quantifying Similarity In Reliability Surfaces Using The Probability Of Agreement, Nathaniel Stevens, C. M. Anderson-Cook

Mathematics

When separate populations exhibit similar reliability as a function of multiple explanatory variables, combining them into a single population is tempting. This can simplify future predictions and reduce uncertainty associated with estimation. However, combining these populations may introduce bias if the underlying relationships are in fact different. The probability of agreement formally and intuitively quantifies the similarity of estimated reliability surfaces across a two-factor input space. An example from the reliability literature demonstrates the utility of the approach when deciding whether to combine two populations or to keep them as distinct. New graphical summaries provide strategies for visualizing the results.


Performance Modeling And Optimization Techniques For Heterogeneous Computing, Supada Laosooksathit Jan 2014

Performance Modeling And Optimization Techniques For Heterogeneous Computing, Supada Laosooksathit

Doctoral Dissertations

Since Graphics Processing Units (CPUs) have increasingly gained popularity amoung non-graphic and computational applications, known as General-Purpose computation on GPU (GPGPU), CPUs have been deployed in many clusters, including the world's fastest supercomputer. However, to make the most efficiency from a GPU system, one should consider both performance and reliability of the system.

This dissertation makes four major contributions. First, the two-level checkpoint/restart protocol that aims to reduce the checkpoint and recovery costs with a latency hiding strategy in a system between a CPU (Central Processing Unit) and a GPU is proposed. The experimental results and analysis reveals some benefits, …


Progressive Reliability Method And Its Application To Offshore Mooring Systems, Mir Emad Mousavi, Paolo Gardoni, Mehdi Maadooliat Nov 2013

Progressive Reliability Method And Its Application To Offshore Mooring Systems, Mir Emad Mousavi, Paolo Gardoni, Mehdi Maadooliat

Mathematics, Statistics and Computer Science Faculty Research and Publications

Assessing the reliability of complex systems (e.g. structures) is essential for a reliability-based optimal design that balances safety and costs of such systems. This paper proposes the Progressive Reliability Method (PRM) for the quantification of the reliability of complex systems. The proposed method is a closed-form solution for calculating the probability of failure. The new method is flexible to the definition of “failure” (i.e., can consider serviceability and ultimate-strength failures) and uses the rules of probability theory to estimate the failure probability of the system or its components. The method is first discussed in general and then illustrated in two …


Optimization In Non-Parametric Survival Analysis And Climate Change Modeling, Iuliana Teodorescu Jan 2013

Optimization In Non-Parametric Survival Analysis And Climate Change Modeling, Iuliana Teodorescu

USF Tampa Graduate Theses and Dissertations

Many of the open problems of current interest in probability and statistics involve complicated data

sets that do not satisfy the strong assumptions of being independent and identically distributed. Often,

the samples are known only empirically, and making assumptions about underlying parametric

distributions is not warranted by the insufficient information available. Under such circumstances,

the usual Fisher or parametric Bayes approaches cannot be used to model the data or make predictions.

However, this situation is quite often encountered in some of the main challenges facing statistical,

data-driven studies of climate change, clinical studies, or financial markets, to name a few. …


Adjusted Empirical Likelihood Models With Estimating Equations For Accelerated Life Tests, Ni Wang, Jye-Chyi Lu, Di Chen, Paul H. Kvam Jan 2011

Adjusted Empirical Likelihood Models With Estimating Equations For Accelerated Life Tests, Ni Wang, Jye-Chyi Lu, Di Chen, Paul H. Kvam

Department of Math & Statistics Faculty Publications

This article proposes an adjusted empirical likelihood estimation (AMELE) method to model and analyze accelerated life testing data. This approach flexibly and rigorously incorporates distribution assumptions and regression structures by estimating equations within a semiparametric estimation framework. An efficient method is provided to compute the empirical likelihood estimates, and asymptotic properties are studied. Real-life examples and numerical studies demonstrate the advantage of the proposed methodology.


Multi-Cause Degradation Path Model: A Case Study On Rubidium Lamp Degradation, Sun Quan, Paul H. Kvam Jan 2011

Multi-Cause Degradation Path Model: A Case Study On Rubidium Lamp Degradation, Sun Quan, Paul H. Kvam

Department of Math & Statistics Faculty Publications

At the core of satellite rubidium standard clocks is the rubidium lamp, which is a critical piece of equipment in a satellite navigation system. There are many challenges in understanding and improving the reliability of the rubidium lamp, including the extensive lifetime requirement and the dearth of samples available for destructive life tests. Experimenters rely on degradation experiments to assess the lifetime distribution of highly reliable products that seem unlikely to fail under the normal stress conditions, because degradation data can provide extra information about product reliability. Based on recent research on the rubidium lamp, this article presents a multi‐cause …


Advancing Assessment Of Quantitative And Scientific Reasoning, Donna L. Sundre, Amy D. Thelk Jul 2010

Advancing Assessment Of Quantitative And Scientific Reasoning, Donna L. Sundre, Amy D. Thelk

Numeracy

Advancing Assessment of Quantitative and Scientific Reasoning is a four-year NSF Project (DUE-0618599) in part designed to evaluate the generalizability of quantitative (QR) and scientific reasoning (SR) assessment instruments created at James Madison University to four other four-year institutions with very distinct missions and student demographics. This article describes the methods, results, and findings we obtained in our studies. More specifically, we describe how to conduct content-alignment exercises in which faculty members map each item from a prospective test to the student learning objectives taught at the institution. Our results indicated that 92-100% of the QR and SR items were …


Fusion Of Sources Of Evidence With Different Importances And Reliabilities, Florentin Smarandache, Jean Dezert, J.M. Tacnet Jan 2010

Fusion Of Sources Of Evidence With Different Importances And Reliabilities, Florentin Smarandache, Jean Dezert, J.M. Tacnet

Branch Mathematics and Statistics Faculty and Staff Publications

This paper presents a new approach for combining sources of evidences with different importances and reliabilities. Usually, the combination of sources of evidences with different reliabilities is done by the classical Shafer’s discounting approach. Therefore, to consider unequal importances of sources, if any, a similar reliability discounting process is generally used, making no difference between the notion of importance and reliability. In fact, in multicriteria decision context, these notions should be clearly distinguished. This paper shows how this can be done and we provide simple examples to show the differences between both solutions for managing importances and reliabilities of sources. …


Quantitative Literacy Assessments: An Introduction To Testing Tests, Dorothy Wallace, Kim Rheinlander, Steven Woloshin, Lisa Schwartz Jun 2009

Quantitative Literacy Assessments: An Introduction To Testing Tests, Dorothy Wallace, Kim Rheinlander, Steven Woloshin, Lisa Schwartz

Numeracy

This paper describes how professional evaluators construct assessment instruments that work properly to measure the right thing. Constructing an assessment tool begins with getting feedback from relevant experts on the content of questions. The tool is developed and refined through comparison with existing instruments, focus groups and cognitive interviews. The final instrument is formally tested for content validity, usability, reliability and construct validity through a variety of statistical measures. This process of construction is illustrated by two examples relevant to quantitative literacy: the Medical Data Interpretation Test and the Math Attitudes Survey.


Maximum Likelihood Estimation Methodology Comparison For The Three-Parameter Weibull Distribution With Applications To Offshore Oil Spills In The Gulf Of Mexico, William V. Harper, Thomas R. James, Ted G. Eschenbach, Leigh Slauson Jan 2008

Maximum Likelihood Estimation Methodology Comparison For The Three-Parameter Weibull Distribution With Applications To Offshore Oil Spills In The Gulf Of Mexico, William V. Harper, Thomas R. James, Ted G. Eschenbach, Leigh Slauson

Mathematics Faculty Scholarship

Maximum Likelihood estimation of the two-parameter Weibull distribution is straightforward; however, there are multiple methods for maximum likelihood estimation of the three-parameter Weibull. This paper presents an evaluation of these methods using four data sets including oil spill data from the Gulf of Mexico. Highlighted are fairly major differences in the estimated parameters between nine statistical packages. A VBA routine has been developed allowing practitioners to implement three-parameter Weibull maximum likelihood estimate within Excel. The code and support documentation are available free at http:faculty.otterbein.edu/WHarper.


Statistical Models For Hot Electron Degradation In Nano-Scaled Mosfet Devices, Suk Joo Bae, Seong-Joon Kim, Way Kuo, Paul H. Kvam Jan 2007

Statistical Models For Hot Electron Degradation In Nano-Scaled Mosfet Devices, Suk Joo Bae, Seong-Joon Kim, Way Kuo, Paul H. Kvam

Department of Math & Statistics Faculty Publications

In a MOS structure, the generation of hot carrier interface states is a critical feature of the item's reliability. On the nano-scale, there are problems with degradation in transconductance, shift in threshold voltage, and decrease in drain current capability. Quantum mechanics has been used to relate this decrease to degradation, and device failure. Although the lifetime, and degradation of a device are typically used to characterize its reliability, in this paper we model the distribution of hot-electron activation energies, which has appeal because it exhibits a two-point discrete mixture of logistic distributions. The logistic mixture presents computational problems that are …


A Nonlinear Random Coefficients Model For Degradation Testing, Suk Joo Bae, Paul H. Kvam Jan 2004

A Nonlinear Random Coefficients Model For Degradation Testing, Suk Joo Bae, Paul H. Kvam

Department of Math & Statistics Faculty Publications

As an alternative to traditional life testing, degradation tests can be effective in assessing product reliability when measurements of degradation leading to failure can be observed. This article presents a degradation model for highly reliable light displays, such as plasma display panels and vacuum fluorescent displays (VFDs). Standard degradation models fail to capture the burn-in characteristics of VFDs, when emitted light actually increases up to a certain point in time before it decreases (or degrades) continuously. Random coefficients are used to model this phenomenon in a nonlinear way, which allows for a nonmonotonic degradation path. In many situations, the relative …


Nonparametric Estimation Of A Distribution Subject To A Stochastic Precedence Constraint, Miguel A. Arcones, Paul H. Kvam, Francisco J. Samaniego Jan 2002

Nonparametric Estimation Of A Distribution Subject To A Stochastic Precedence Constraint, Miguel A. Arcones, Paul H. Kvam, Francisco J. Samaniego

Department of Math & Statistics Faculty Publications

For any two random variables X and Y with distributions F and G defined on [0,∞), X is said to stochastically precede Y if P(XY) ≥ 1/2. For independent X and Y, stochastic precedence (denoted by XspY) is equivalent to E[G(X–)] ≤ 1/2. The applicability of stochastic precedence in various statistical contexts, including reliability modeling, tests for distributional equality versus various alternatives, and the relative performance of comparable tolerance bounds, is discussed. The problem of estimating the underlying distribution(s) of experimental data under the assumption that they obey a …


Average-Case Lower Bounds For Noisy Boolean Decision Trees, William Evans, Nicholas Pippenger Jan 1998

Average-Case Lower Bounds For Noisy Boolean Decision Trees, William Evans, Nicholas Pippenger

All HMC Faculty Publications and Research

We present a new method for deriving lower bounds to the expected number of queries made by noisy decision trees computing Boolean functions. The new method has the feature that expectations are taken with respect to a uniformly distributed random input, as well as with respect to the random noise, thus yielding stronger lower bounds. It also applies to many more functions than do previous results. The method yields a simple proof of the result (previously established by Reischuk and Schmeltz) that almost all Boolean functions of n arguments require $\Me(n \log n)$ queries, and strengthens this bound from the …