Open Access. Powered by Scholars. Published by Universities.®
- Institution
-
- University of Nebraska - Lincoln (9)
- Florida International University (8)
- St. Cloud State University (7)
- COBRA (5)
- Kennesaw State University (5)
-
- Technological University Dublin (5)
- University of Central Florida (4)
- City University of New York (CUNY) (3)
- The British University in Egypt (3)
- University of Connecticut (3)
- University of Massachusetts Amherst (2)
- Wayne State University (2)
- California Polytechnic State University, San Luis Obispo (1)
- Chapman University (1)
- Lawrence University (1)
- Liberty University (1)
- Munster Technological University (1)
- Old Dominion University (1)
- Syracuse University (1)
- University of Arkansas, Fayetteville (1)
- University of Kentucky (1)
- University of New Mexico (1)
- University of Richmond (1)
- University of Texas Rio Grande Valley (1)
- Western Kentucky University (1)
- Keyword
-
- Statistics (5)
- Data Analysis (4)
- K-12 Education (4)
- Standardized Tests (4)
- Community Corrections (3)
-
- Monte Carlo Simulation (3)
- Regression (3)
- AUC (2)
- Baseball (2)
- Correlation (2)
- EM algorithm (2)
- Feature selection (2)
- Logistic regression (2)
- Meta-regression (2)
- Poisson (2)
- Risk modeling (2)
- Simulation (2)
- Sports (2)
- 19th century literature (1)
- 3.3 HEALTH SCIENCES (1)
- ABX index (1)
- Accelerated degradation data (1)
- Accuracy (1)
- Agricultural drought risk (1)
- Air Pollution (1)
- Algorithm (1)
- Analytics (1)
- Antimicrobial Resistance (1)
- Applied mathematics (1)
- Artificial Intelligence (1)
- Publication Year
- Publication
-
- FIU Electronic Theses and Dissertations (8)
- Published and Grey Literature from PhD Candidates (5)
- Data Science and Data Mining (4)
- School District Data (4)
- Basic Science Engineering (3)
-
- COBRA Preprint Series (3)
- Criminal Justice Data (3)
- Department of Statistics: Dissertations, Theses, and Student Work (3)
- Publications and Research (3)
- Articles (2)
- CHIP Documents (2)
- Conference papers (2)
- Kinesiology, Health and Sport Studies (2)
- Masters Theses 1911 - February 2014 (2)
- Books/Book chapters (1)
- Branch Mathematics and Statistics Faculty and Staff Publications (1)
- Community & Environmental Health Faculty Publications (1)
- Department of Animal Science: Faculty Publications (1)
- Department of Civil and Environmental Engineering: Dissertations, Theses, and Student Research (1)
- Department of English: Faculty Publications (1)
- Department of Management: Faculty Publications (1)
- Department of Math & Statistics Faculty Publications (1)
- Department of Mathematics Publications (1)
- Department of Statistics: Faculty Publications (1)
- English Language Institute (1)
- Honors Scholar Theses (1)
- Lawrence University Honors Projects (1)
- Lewis Honors College Capstone Collection (1)
- Masters Theses & Specialist Projects (1)
- Mathematical Sciences Spring Lecture Series (1)
Articles 1 - 30 of 69
Full-Text Articles in Applied Statistics
Predicting Superconducting Critical Temperature Using Regression Analysis, Roland Fiagbe
Predicting Superconducting Critical Temperature Using Regression Analysis, Roland Fiagbe
Data Science and Data Mining
This project estimates a regression model to predict the superconducting critical temperature based on variables extracted from the superconductor’s chemical formula. The regression model along with the stepwise variable selection gives a reasonable and good predictive model with a lower prediction error (MSE). Variables extracted based on atomic radius, valence, atomic mass and thermal conductivity appeared to have the most contribution to the predictive model.
Machine Learning Approaches For Cyberbullying Detection, Roland Fiagbe
Machine Learning Approaches For Cyberbullying Detection, Roland Fiagbe
Data Science and Data Mining
Cyberbullying refers to the act of bullying using electronic means and the internet. In recent years, this act has been identifed to be a major problem among young people and even adults. It can negatively impact one’s emotions and lead to adverse outcomes like depression, anxiety, harassment, and suicide, among others. This has led to the need to employ machine learning techniques to automatically detect cyberbullying and prevent them on various social media platforms. In this study, we want to analyze the combination of some Natural Language Processing (NLP) algorithms (such as Bag-of-Words and TFIDF) with some popular machine learning …
Movie Recommender System Using Matrix Factorization, Roland Fiagbe
Movie Recommender System Using Matrix Factorization, Roland Fiagbe
Data Science and Data Mining
Recommendation systems are a popular and beneficial field that can help people make informed decisions automatically. This technique assists users in selecting relevant information from an overwhelming amount of available data. When it comes to movie recommendations, two common methods are collaborative filtering, which compares similarities between users, and content-based filtering, which takes a user’s specific preferences into account. However, our study focuses on the collaborative filtering approach, specifically matrix factorization. Various similarity metrics are used to identify user similarities for recommendation purposes. Our project aims to predict movie ratings for unwatched movies using the MovieLens rating dataset. We developed …
Small But Mighty: Examing The Utility Of Microstatistics In Modeling Ice Hockey, Matt Palmer
Small But Mighty: Examing The Utility Of Microstatistics In Modeling Ice Hockey, Matt Palmer
Senior Honors Theses
As research into hockey analytics continues, an increasing number of metrics are being introduced into the knowledge base of the field, creating a need to determine whether various stats are useful or simply add noise to the discussion. This paper examines microstatistics – manually tracked metrics which go beyond the NHL’s publicly released stats – both through the lens of meta-analytics (which attempt to objectively assess how useful a metric is) and modeling game probabilities. Results show that while there is certainly room for improvement in understanding and use of microstats in modeling, the metrics overall represent an area of …
Uconn Baseball Batting Order Optimization, Gavin Rublewski, Gavin Rublewski
Uconn Baseball Batting Order Optimization, Gavin Rublewski, Gavin Rublewski
Honors Scholar Theses
Challenging conventional wisdom is at the very core of baseball analytics. Using data and statistical analysis, the sets of rules by which coaches make decisions can be justified, or possibly refuted. One of those sets of rules relates to the construction of a batting order. Through data collection, data adjustment, the construction of a baseball simulator, and the use of a Monte Carlo Simulation, I have assessed thousands of possible batting orders to determine the roster-specific strategies that lead to optimal run production for the 2023 UConn baseball team. This paper details a repeatable process in which basic player statistics …
Classification Of Adult Income Using Decision Tree, Roland Fiagbe
Classification Of Adult Income Using Decision Tree, Roland Fiagbe
Data Science and Data Mining
Decision tree is a commonly used data mining methodology for performing classification tasks. It is a tree-based supervised machine learning algorithm that is used to classify or make predictions in a path of how previous questions are answered. Generally, the decision tree algorithm categorizes data into branch-like segments that develop into a tree that contains a root, nodes, and leaves. This project seeks to explore the decision tree methodology and apply it to the Adult Income dataset from the UCI Machine Learning Repository, to determine whether a person makes over 50K per year and determine the necessary factors that improve …
Forecasting Remission Time Of A Treatment Method For Leukemia As An Application To Statistical Inference Approach, Ahmed Galal Atia, Mahmoud Mansour, Rashad Mohamed El-Sagheer, B. S. El-Desouky
Forecasting Remission Time Of A Treatment Method For Leukemia As An Application To Statistical Inference Approach, Ahmed Galal Atia, Mahmoud Mansour, Rashad Mohamed El-Sagheer, B. S. El-Desouky
Basic Science Engineering
In this paper, Weibull-Linear Exponential distribution (WLED) has been investigated whether being it is a well-fit distribution to a clinical real data. These data represent the duration of remission achieved by a certain drug used in the treatment of leukemia for a group of patients. The statistical inference approach is used to estimate the parameters of the WLED through the set of the fitted data. The estimated parameters are utilized to evaluate the survival and hazard functions and hence assessing the treatment method through forecasting the duration of remission times of patients. A two-sample prediction approach has been applied to …
A Bayesian Programming Approach To Car-Following Model Calibration And Validation Using Limited Data, Franklin Abodo
A Bayesian Programming Approach To Car-Following Model Calibration And Validation Using Limited Data, Franklin Abodo
FIU Electronic Theses and Dissertations
Traffic simulation software is used by transportation researchers and engineers to design and evaluate changes to roadway networks. Underlying these simulators are mathematical models of microscopic driver behavior from which macroscopic measures of flow and congestion can be recovered. Many models are intended to apply to only a subset of possible traffic scenarios and roadway configurations, while others do not have any explicit constraint on their applicability. Work zones on highways are one scenario for which no model invented to date has been shown to accurately reproduce realistic driving behavior. This makes it difficult to optimize for safety and other …
A Monte Carlo Analysis Of Seven Dichotomous Variable Confidence Interval Equations, Morgan Juanita Dubose
A Monte Carlo Analysis Of Seven Dichotomous Variable Confidence Interval Equations, Morgan Juanita Dubose
Masters Theses & Specialist Projects
Department of Psychological Sciences Western Kentucky University There are two options to estimate a range of likely values for the population mean of a continuous variable: one for when the population standard deviation is known and another for when the population standard deviation is unknown. There are seven proposed equations to calculate the confidence interval for the population mean of a dichotomous variable: normal approximation interval, Wilson interval, Jeffreys interval, Clopper-Pearson, Agresti-Coull, arcsine transformation, and logit transformation. In this study, I compared the percent effectiveness of each equation using a Monte Carlo analysis and the interval range over a range …
A Simple Algorithm For Generating A New Two Sample Type-Ii Progressive Censoring With Applications, E. M. Shokr, Rashad Mohamed El-Sagheer, Mahmoud Mansour, H. M. Faied, B. S. El-Desouky
A Simple Algorithm For Generating A New Two Sample Type-Ii Progressive Censoring With Applications, E. M. Shokr, Rashad Mohamed El-Sagheer, Mahmoud Mansour, H. M. Faied, B. S. El-Desouky
Basic Science Engineering
In this article, we introduce a simple algorithm to generating a new type-II progressive censoring scheme for two samples. It is observed that the proposed algorithm can be applied for any continues probability distribution. Moreover, the description model and necessary assumptions are discussed. In addition, the steps of simple generation algorithm along with programming steps are also constructed on real example. The inference of two Weibull Frechet populations are discussed under the proposed algorithm. Both classical and Bayesian inferential approaches of the distribution parameters are discussed. Furthermore, approximate confidence intervals are constructed based on the asymptotic distribution of the maximum …
Application Of Randomness In Finance, Jose Sanchez, Daanial Ahmad, Satyanand Singh
Application Of Randomness In Finance, Jose Sanchez, Daanial Ahmad, Satyanand Singh
Publications and Research
Brownian Motion which is also considered to be a Wiener process and can be thought of as a random walk. In our project we had briefly discussed the fluctuations of financial indices and related it to Brownian Motion and the modeling of Stock prices.
Lecture 04: Spatial Statistics Applications Of Hrl, Trl, And Mixed Precision, David Keyes
Lecture 04: Spatial Statistics Applications Of Hrl, Trl, And Mixed Precision, David Keyes
Mathematical Sciences Spring Lecture Series
As simulation and analytics enter the exascale era, numerical algorithms, particularly implicit solvers that couple vast numbers of degrees of freedom, must span a widening gap between ambitious applications and austere architectures to support them. We present fifteen universals for researchers in scalable solvers: imperatives from computer architecture that scalable solvers must respect, strategies towards achieving them that are currently well established, and additional strategies currently being developed for an effective and efficient exascale software ecosystem. We consider recent generalizations of what it means to “solve” a computational problem, which suggest that we have often been “oversolving” them at the …
On The Evolution Equation For Modelling The Covid-19 Pandemic, Jonathan Blackledge
On The Evolution Equation For Modelling The Covid-19 Pandemic, Jonathan Blackledge
Books/Book chapters
The paper introduces and discusses the evolution equation, and, based exclusively on this equation, considers random walk models for the time series available on the daily confirmed Covid-19 cases for different countries. It is shown that a conventional random walk model is not consistent with the current global pandemic time series data, which exhibits non-ergodic properties. A self-affine random walk field model is investigated, derived from the evolutionary equation for a specified memory function which provides the non-ergodic fields evident in the available Covid-19 data. This is based on using a spectral scaling relationship of the type 1/ωα where ω …
Interval Estimation Of Proportion Of Second-Level Variance In Multi-Level Modeling, Steven Svoboda
Interval Estimation Of Proportion Of Second-Level Variance In Multi-Level Modeling, Steven Svoboda
The Nebraska Educator: A Student-Led Journal
Physical, behavioral and psychological research questions often relate to hierarchical data systems. Examples of hierarchical data systems include repeated measures of students nested within classrooms, nested within schools and employees nested within supervisors, nested within organizations. Applied researchers studying hierarchical data structures should have an estimate of the intraclass correlation coefficient (ICC) for every nested level in their analyses because ignoring even relatively small amounts of interdependence is known to inflate Type I error rate in single-level models. Traditionally, researchers rely upon the ICC as a point estimate of the amount of interdependency in their data. Recent methods utilizing an …
Statistical Methodology To Establish A Benchmark For Evaluating Antimicrobial Resistance Genes Through Real Time Pcr Assay, Enakshy Dutta
Statistical Methodology To Establish A Benchmark For Evaluating Antimicrobial Resistance Genes Through Real Time Pcr Assay, Enakshy Dutta
Department of Statistics: Dissertations, Theses, and Student Work
Novel diagnostic tests are usually compared with gold standard tests for evaluating diagnostic accuracy. For assessing antimicrobial resistance (AMR) to bovine respiratory disease (BRD) pathogens, phenotypic broth microdilution method is used as gold standard (GS). The objective of the thesis is to evaluate the optimal cycle threshold (Ct) generated by real-time polymerase chain reaction (rtPCR) to genes that confer resistance that will translate to the phenotypic classification of AMR. Data from two different methodologies are assessed to identify Ct that will discriminate between resistance (R) and susceptibility (S). First, the receiver operating characteristic (ROC) curve was used to determine the …
Working Children On Java Island 2017, Yuniarti
Working Children On Java Island 2017, Yuniarti
English Language Institute
Children's wellbeing has currently become a global concern as many of them are engaged in the labor force. A small area estimation (SAE) technique, EBLUP under Fey Herriot model, is employed to reveal their number in regencies of Java Island. Statistics have been disaggregated by geographical location (urban/rural) and gender. These statistics are required by the government as the basis for policy making.
An Automatic Interaction Detection Hybrid Model For Bankcard Response Classification, Yan Wang, Sherry Ni, Brian Stone
An Automatic Interaction Detection Hybrid Model For Bankcard Response Classification, Yan Wang, Sherry Ni, Brian Stone
Published and Grey Literature from PhD Candidates
Data mining techniques have numerous applications in bankcard response modeling. Logistic regression has been used as the standard modeling tool in the financial industry because of its almost always desirable performance and its interpretability. In this paper, we propose a hybrid bankcard response model, which integrates decision tree-based chi-square automatic interaction detection (CHAID) into logistic regression. In the first stage of the hybrid model, CHAID analysis is used to detect the possible potential variable interactions. Then in the second stage, these potential interactions are served as the additional input variables in logistic regression. The motivation of the proposed hybrid model …
A Two-Stage Hybrid Model By Using Artificial Neural Networks As Feature Construction Algorithms, Yan Wang, Sherry Ni, Brian Stone
A Two-Stage Hybrid Model By Using Artificial Neural Networks As Feature Construction Algorithms, Yan Wang, Sherry Ni, Brian Stone
Published and Grey Literature from PhD Candidates
We propose a two-stage hybrid approach with neural networks as the new feature construction algorithms for bankcard response classifications. The hybrid model uses a very simple neural network structure as the new feature construction tool in the first stage, then the newly created features are used as the additional input variables in logistic regression in the second stage. The model is compared with the traditional one-stage model in credit customer response classification. It is observed that the proposed two-stage model outperforms the one-stage model in terms of accuracy, the area under the ROC curve, and KS statistic. By creating new …
Predicting Class-Imbalanced Business Risk Using Resampling, Regularization, And Model Ensembling Algorithms, Yan Wang, Sherry Ni
Predicting Class-Imbalanced Business Risk Using Resampling, Regularization, And Model Ensembling Algorithms, Yan Wang, Sherry Ni
Published and Grey Literature from PhD Candidates
We aim at developing and improving the imbalanced business risk modeling via jointly using proper evaluation criteria, resampling, cross-validation, classifier regularization, and ensembling techniques. Area Under the Receiver Operating Characteristic Curve (AUC of ROC) is used for model comparison based on 10-fold cross-validation. Two undersampling strategies including random undersampling (RUS) and cluster centroid undersampling (CCUS), as well as two oversampling methods including random oversampling (ROS) and Synthetic Minority Oversampling Technique (SMOTE), are applied. Three highly interpretable classifiers, including logistic regression without regularization (LR), L1-regularized LR (L1LR), and decision tree (DT) are implemented. Two ensembling techniques, including Bagging and Boosting, are …
A Xgboost Risk Model Via Feature Selection And Bayesian Hyper-Parameter Optimization, Yan Wang, Sherry Ni
A Xgboost Risk Model Via Feature Selection And Bayesian Hyper-Parameter Optimization, Yan Wang, Sherry Ni
Published and Grey Literature from PhD Candidates
This paper aims to explore models based on the extreme gradient boosting (XGBoost) approach for business risk classification. Feature selection (FS) algorithms and hyper-parameter optimizations are simultaneously considered during model training. The five most commonly used FS methods including weight by Gini, weight by Chi-square, hierarchical variable clustering, weight by correlation, and weight by information are applied to alleviate the effect of redundant features. Two hyper-parameter optimization approaches, random search (RS) and Bayesian tree-structuredParzen Estimator (TPE), are applied in XGBoost. The effect of different FS and hyper-parameter optimization methods on the model performance are investigated by the Wilcoxon Signed Rank …
Modelling Interactions Among Offenders: A Latent Space Approach For Interdependent Ego-Networks, Isabella Gollini, Alberto Caimo, Paolo Campana
Modelling Interactions Among Offenders: A Latent Space Approach For Interdependent Ego-Networks, Isabella Gollini, Alberto Caimo, Paolo Campana
Articles
Illegal markets are notoriously difficult to study. Police data offer an increasingly exploited source of evidence. However, their secondary nature poses challenges for researchers. A key issue is that researchers often have to deal with two sets of actors: targeted and non-targeted. This work develops a latent space model for interdependent ego-networks purposely created to deal with the targeted nature of police evidence. By treating targeted offenders as egos and their contacts as alters, the model (a) leverages on the full information available and (b) mirrors the specificity of the data collection strategy. The paper then applies this approach to …
Sex And Age Differences In Prevalence And Risk Factors For Prediabetes In Mexican-Americans, Kristina Vatcheva, Belinda M. Reininger, Susan P. Fisher-Hoch, Joseph B. Mccormick
Sex And Age Differences In Prevalence And Risk Factors For Prediabetes In Mexican-Americans, Kristina Vatcheva, Belinda M. Reininger, Susan P. Fisher-Hoch, Joseph B. Mccormick
School of Mathematical and Statistical Sciences Faculty Publications and Presentations
AIMS:
Over 1/3 of Americans have prediabetes, while 9.4% have type 2 diabetes. The aim of our study was to estimate the prevalence of prediabetes in Mexican Americans, with known 28.2% prevalence of type 2 diabetes, by age and sex and to identify critical socio-demographic and clinical factors associated with prediabetes.
METHODS:
Data were collected between 2004 and 2017 from the Cameron County Hispanic Cohort in Texas. Weighted crude and sex- and age- stratified prevalences were calculated. Survey weighted logistic regression analyses were conducted to identify risk factors for prediabetes.
RESULTS:
The prevalence of prediabetes (32%) was slightly higher than …
Best Probable Subset: A New Method For Reducing Data Dimensionality In Linear Regression, Elieser Nodarse
Best Probable Subset: A New Method For Reducing Data Dimensionality In Linear Regression, Elieser Nodarse
FIU Electronic Theses and Dissertations
Regression is a statistical technique for modeling the relationship between a dependent variable Y and two or more predictor variables, also known as regressors. In the broad field of regression, there exists a special case in which the relationship between the dependent variable and the regressor(s) is linear. This is known as linear regression.
The purpose of this paper is to create a useful method that effectively selects a subset of regressors when dealing with high dimensional data and/or collinearity in linear regression. As the name depicts it, high dimensional data occurs when the number of predictor variables is far …
Computational Analysis Of Large-Scale Trends And Dynamics In Eukaryotic Protein Family Evolution, Joseph Boehm Ahrens
Computational Analysis Of Large-Scale Trends And Dynamics In Eukaryotic Protein Family Evolution, Joseph Boehm Ahrens
FIU Electronic Theses and Dissertations
The myriad protein-coding genes found in present-day eukaryotes arose from a combination of speciation and gene duplication events, spanning more than one billion years of evolution. Notably, as these proteins evolved, the individual residues at each site in their amino acid sequences were replaced at markedly different rates. The relationship between protein structure, protein function, and site-specific rates of amino acid replacement is a topic of ongoing research. Additionally, there is much interest in the different evolutionary constraints imposed on sequences related by speciation (orthologs) versus sequences related by gene duplication (paralogs). A principal aim of this dissertation is to …
Unified Methods For Feature Selection In Large-Scale Genomic Studies With Censored Survival Outcomes, Lauren Spirko-Burns, Karthik Devarajan
Unified Methods For Feature Selection In Large-Scale Genomic Studies With Censored Survival Outcomes, Lauren Spirko-Burns, Karthik Devarajan
COBRA Preprint Series
One of the major goals in large-scale genomic studies is to identify genes with a prognostic impact on time-to-event outcomes which provide insight into the disease's process. With rapid developments in high-throughput genomic technologies in the past two decades, the scientific community is able to monitor the expression levels of tens of thousands of genes and proteins resulting in enormous data sets where the number of genomic features is far greater than the number of subjects. Methods based on univariate Cox regression are often used to select genomic features related to survival outcome; however, the Cox model assumes proportional hazards …
Controlling For Confounding Via Propensity Score Methods Can Result In Biased Estimation Of The Conditional Auc: A Simulation Study, Hadiza I. Galadima, Donna K. Mcclish
Controlling For Confounding Via Propensity Score Methods Can Result In Biased Estimation Of The Conditional Auc: A Simulation Study, Hadiza I. Galadima, Donna K. Mcclish
Community & Environmental Health Faculty Publications
In the medical literature, there has been an increased interest in evaluating association between exposure and outcomes using nonrandomized observational studies. However, because assignments to exposure are not random in observational studies, comparisons of outcomes between exposed and nonexposed subjects must account for the effect of confounders. Propensity score methods have been widely used to control for confounding, when estimating exposure effect. Previous studies have shown that conditioning on the propensity score results in biased estimation of conditional odds ratio and hazard ratio. However, research is lacking on the performance of propensity score methods for covariate adjustment when estimating the …
Statistical Investigation Of Road And Railway Hazardous Materials Transportation Safety, Amirfarrokh Iranitalab
Statistical Investigation Of Road And Railway Hazardous Materials Transportation Safety, Amirfarrokh Iranitalab
Department of Civil and Environmental Engineering: Dissertations, Theses, and Student Research
Transportation of hazardous materials (hazmat) in the United States (U.S.) constituted 22.8% of the total tonnage transported in 2012 with an estimated value of more than 2.3 billion dollars. As such, hazmat transportation is a significant economic activity in the U.S. However, hazmat transportation exposes people and environment to the infrequent but potentially severe consequences of incidents resulting in hazmat release. Trucks and trains carried 63.7% of the hazmat in the U.S. in 2012 and are the major foci of this dissertation. The main research objectives were 1) identification and quantification of the effects of different factors on occurrence and …
Season-Ahead Forecasting Of Water Storage And Irrigation Requirements – An Application To The Southwest Monsoon In India, Arun Ravindranath, Naresh Devineni, Upmanu Lall, Paulina Concha Larrauri
Season-Ahead Forecasting Of Water Storage And Irrigation Requirements – An Application To The Southwest Monsoon In India, Arun Ravindranath, Naresh Devineni, Upmanu Lall, Paulina Concha Larrauri
Publications and Research
Water risk management is a ubiquitous challenge faced by stakeholders in the water or agricultural sector. We present a methodological framework for forecasting water storage requirements and present an application of this methodology to risk assessment in India. The application focused on forecasting crop water stress for potatoes grown during the monsoon season in the Satara district of Maharashtra. Pre-season large-scale climate predictors used to forecast water stress were selected based on an exhaustive search method that evaluates for highest ranked probability skill score and lowest root-mean-squared error in a leave-one-out cross-validation mode. Adaptive forecasts were made in the years …
Motivational Interviewing Treatment Integrity (Miti), Himal Adhikari
Motivational Interviewing Treatment Integrity (Miti), Himal Adhikari
Criminal Justice Data
This report summarizes interview data collected using MITI, by the Stearns County Community Corrections. The 42 interviews of corrections clients were held between 2015 and 2017. They were examined to determine the effectiveness of interviewing techniques using criteria established on the MITI form. Overall effectiveness measures are given, as well as correlations among some of the quantitative measurements.
Reincarceration Rates For Dwi And Domestic Violence Offenders, Torrence Savage, Brittany Wittrock, Joseph Brough
Reincarceration Rates For Dwi And Domestic Violence Offenders, Torrence Savage, Brittany Wittrock, Joseph Brough
Criminal Justice Data
Stearns County Community Corrections is looking for the likelihood for offenders to be re-incarcerated within ninety days to three years after the last contact between the agent and the client. All of the clients were separated by year and categorized by whether their initial crime was Domestic Violence or DWI. Many of these clients had multiple observations for cases that were reduced to include only the case with the agent that they worked with the longest.
It was found that the re-incarceration rate for both Domestic and DWI had a decrease of approximately 75% over the course of four years. …