Open Access. Powered by Scholars. Published by Universities.®

Engineering Commons

Open Access. Powered by Scholars. Published by Universities.®

California Polytechnic State University, San Luis Obispo

Computer Vision

Articles 1 - 21 of 21

Full-Text Articles in Engineering

Insights Into Cellular Evolution: Temporal Deep Learning Models And Analysis For Cell Image Classification, Xinran Zhao Mar 2024

Insights Into Cellular Evolution: Temporal Deep Learning Models And Analysis For Cell Image Classification, Xinran Zhao

Master's Theses

Understanding the temporal evolution of cells poses a significant challenge in developmental biology. This study embarks on a comparative analysis of various machine-learning techniques to classify cell colony images across different timestamps, thereby aiming to capture dynamic transitions of cellular states. By performing Transfer Learning with state-of-the-art classification networks, we achieve high accuracy in categorizing single-timestamp images. Furthermore, this research introduces the integration of temporal models, notably LSTM (Long Short Term Memory Network), R-Transformer (Recurrent Neural Network enhanced Transformer) and ViViT (Video Vision Transformer), to undertake this classification task to verify the effectiveness of incorporating temporal features into the classification …


Smartphone Based Object Detection For Shark Spotting, Darrick W. Oliver Nov 2023

Smartphone Based Object Detection For Shark Spotting, Darrick W. Oliver

Master's Theses

Given concern over shark attacks in coastal regions, the recent use of unmanned aerial vehicles (UAVs), or drones, has increased to ensure the safety of beachgoers. However, much of city officials' process remains manual, with drone operation and review of footage still playing a significant role. In pursuit of a more automated solution, researchers have turned to the usage of neural networks to perform detection of sharks and other marine life. For on-device solutions, this has historically required assembling individual hardware components to form an embedded system to utilize the machine learning model. This means that the camera, neural processing …


Hardware-In-The-Loop Reaction Wheel Testbed With Camera Vision, Abigail Romero, Harvey Perkins, Stephen Kwok-Choon Oct 2023

Hardware-In-The-Loop Reaction Wheel Testbed With Camera Vision, Abigail Romero, Harvey Perkins, Stephen Kwok-Choon

College of Engineering Summer Undergraduate Research Program

Reaction wheels are widely used in aerospace systems as a method of attitude control. This research was focused on the design, development, and testing of a hardware-in-the-loop reaction wheel testbed that can be used for research and teaching applications related to satellite navigation and control. This project successfully utilized commercial off-the-shelf components to develop a reaction wheel capable of controlling the orientation of a freely rotating platform, as well as tracking objects using computer vision.


Experimental Characterization And Computer Vision-Assisted Detection Of Pitting Corrosion On Stainless Steel Structural Members, Riley J. Muehler Jun 2023

Experimental Characterization And Computer Vision-Assisted Detection Of Pitting Corrosion On Stainless Steel Structural Members, Riley J. Muehler

Master's Theses

Pitting corrosion is a prevalent form of corrosive damage that can weaken, damage, and initiate failure in corrosion-resistant metallic materials. For instance, 304 stainless steel is commonly utilized in various structures (e.g., miter gates, heat exchangers, and storage tanks), but is prone to failure through pitting corrosion and stress corrosion cracking under mechanical loading, regardless of its high corrosion resistance. In this study, to better understand the pitting corrosion damage development, controlled corrosion experiments were conducted to generate pits on 304 stainless steel specimens with and without mechanical loading. The pit development over time was characterized using a high-resolution laser …


Iris Detection Authenticator, Nathan D. Tang, Bryan K. Chau Nov 2022

Iris Detection Authenticator, Nathan D. Tang, Bryan K. Chau

Electrical Engineering

The development of iris biometric identification recognition is presented. Iris recognition differs from other methods because data acquisition is non-physical and is more accessible. It has been proven that the iris does not change as an individual ages and is well protected from external damages due to the eyelid and cornea, acting as a shield to the iris. In addition, the iris is almost impossible to forge due to its complex patterns and the current limitations in technology. Using Canny Edge Detection, Hough Transform, rubber-sheet normalization, Histogram of Gradient feature extraction, and the MultiMedia University iris database as our subjects, …


Eagle Medical Tray Denesting & Debris Removal Process, Nicholas Allen Ungefug, Noah Chavez, Susana Shu-Lin Okhuysen, Michael Augustine Pennington Jun 2022

Eagle Medical Tray Denesting & Debris Removal Process, Nicholas Allen Ungefug, Noah Chavez, Susana Shu-Lin Okhuysen, Michael Augustine Pennington

Industrial and Manufacturing Engineering

Eagle Medical Incorporated is a contract medical device packaging and sterilization company. The company purchases thermoformed medical packaging trays, which maintain the sterility of medical devices, from various manufacturers. To ensure packaging quality and to prevent cleanroom contamination, Eagle Medical inspects and sterilizes each blister tray that they order. This process is an essential non-value-added activity that creates a bottleneck. Cleanroom employees must stop packaging medical devices and attend to the processing of blister trays and packaging solutions. The blister trays arrive at Eagle’s facility in nested stacks. Vibration and movement during shipping further compresses the stacks, which makes separation …


An Analysis Of Camera Configurations And Depth Estimation Algorithms For Triple-Camera Computer Vision Systems, Jared Peter-Contesse Dec 2021

An Analysis Of Camera Configurations And Depth Estimation Algorithms For Triple-Camera Computer Vision Systems, Jared Peter-Contesse

Master's Theses

The ability to accurately map and localize relevant objects surrounding a vehicle is an important task for autonomous vehicle systems. Currently, many of the environmental mapping approaches rely on the expensive LiDAR sensor. Researchers have been attempting to transition to cheaper sensors like the camera, but so far, the mapping accuracy of single-camera and dual-camera systems has not matched the accuracy of LiDAR systems. This thesis examines depth estimation algorithms and camera configurations of a triple-camera system to determine if sensor data from an additional perspective will improve the accuracy of camera-based systems. Using a synthetic dataset, the performance of …


Visual Speech Recognition Using A 3d Convolutional Neural Network, Matthew Rochford Dec 2019

Visual Speech Recognition Using A 3d Convolutional Neural Network, Matthew Rochford

Master's Theses

Main stream automatic speech recognition (ASR) makes use of audio data to identify spoken words, however visual speech recognition (VSR) has recently been of increased interest to researchers. VSR is used when audio data is corrupted or missing entirely and also to further enhance the accuracy of audio-based ASR systems. In this research, we present both a framework for building 3D feature cubes of lip data from videos and a 3D convolutional neural network (CNN) architecture for performing classification on a dataset of 100 spoken words, recorded in an uncontrolled envi- ronment. Our 3D-CNN architecture achieves a testing accuracy of …


Strawberry Detection Under Various Harvestation Stages, Yavisht Fitter Mar 2019

Strawberry Detection Under Various Harvestation Stages, Yavisht Fitter

Master's Theses

This paper analyzes three techniques attempting to detect strawberries at various stages in its growth cycle. Histogram of Oriented Gradients (HOG), Local Binary Patterns (LBP) and Convolutional Neural Networks (CNN) were implemented on a limited custom-built dataset. The methodologies were compared in terms of accuracy and computational efficiency. Computational efficiency is defined in terms of image resolution as testing on a smaller dimensional image is much quicker than larger dimensions. The CNN based implementation obtained the best results with an 88% accuracy at the highest level of efficiency as well (600x800). LBP generated moderate results with a 74% detection accuracy …


Localization Using Convolutional Neural Networks, Shannon D. Fong Dec 2018

Localization Using Convolutional Neural Networks, Shannon D. Fong

Computer Engineering

With the increased accessibility to powerful GPUs, ability to develop machine learning algorithms has increased significantly. Coupled with open source deep learning frameworks, average users are now able to experiment with convolutional neural networks (CNNs) to solve novel problems. This project sought to train a CNN capable of classifying between various locations within a building. A single continuous video was taken while standing at each desired location so that every class in the neural network was represented by a single video. Each location was given a number to be used for classification and the video was subsequently titled locX. These …


Lionfish Detection System, Carmelo Furlan, Andrew Boniface Jun 2018

Lionfish Detection System, Carmelo Furlan, Andrew Boniface

Computer Engineering

Deep neural networks have proven to be an effective method in classification of images. The ability to recognize objects has opened the door for many new systems which use image classification to solve challenging problems where conventional image classification would be inadequate. We trained a large, deep convolutional neural network to identify lionfish from other species that might be found in the same habitats. Google’s Inception framework served as a powerful platform for our fish recognition system. By using transfer learning, we were able to obtain exceptional results for the classification of different species of fish. The convolutional neural network …


Corridor Navigation For Monocular Vision Mobile Robots, Matthew James Ng Jun 2018

Corridor Navigation For Monocular Vision Mobile Robots, Matthew James Ng

Master's Theses

Monocular vision robots use a single camera to process information about its environment. By analyzing this scene, the robot can determine the best navigation direction. Many modern approaches to robot hallway navigation involve using a plethora of sensors to detect certain features in the environment. This can be laser range finders, inertial measurement units, motor encoders, and cameras.

By combining all these sensors, there is unused data which could be useful for navigation. To draw back and develop a baseline approach, this thesis explores the reliability and capability of solely using a camera for navigation. The basic navigation structure begins …


Automated Pruning Of Greenhouse Indeterminate Tomato Plants, Joey M. Angeja Jun 2018

Automated Pruning Of Greenhouse Indeterminate Tomato Plants, Joey M. Angeja

Master's Theses

Pruning of indeterminate tomato plants is vital for a profitable yield and it still remains a manual process. There has been research in automated pruning of grapevines, trees, and other plants, but tomato plants have yet to be explored. Wage increases are contributing to the depleting profits of greenhouse tomato farmers. Rises in population are the driving force behind the need for efficient growing techniques. The major contribution of this thesis is a computer vision algorithm for detecting greenhouse tomato pruning points without the use of depth sensors. Given an up-close 2-D image of a tomato stem with the background …


Underwater Computer Vision - Fish Recognition, Spencer Chang, Austin Otto Jun 2017

Underwater Computer Vision - Fish Recognition, Spencer Chang, Austin Otto

Computer Engineering

The Underwater Computer Vision – Fish Recognition project includes the design and implementation of a device that can withstand staying underwater for a duration of time, take pictures of underwater creatures, such as fish, and be able to identify certain fish. The system is meant to be cheap to create, yet still able to process the images it takes and identify the objects in the pictures with some accuracy. The device can output its results to another device or an end user.


Multispectral Identification Array, Zachary D. Eagan Jun 2017

Multispectral Identification Array, Zachary D. Eagan

Computer Engineering

The Multispectral Identification Array is a device for taking full image spectroscopy data via the illumination of a subject with sixty-four unique spectra. The array combines images under the illumination spectra to produce an approximate reflectance graph for every pixel in a scene. Acquisition of an entire spectrum allows the array to differentiate objects based on surface material. Spectral graphs produced are highly approximate and should not be used to determine material properties, however the output is sufficiently consistent to allow differentiation and identification of previously sampled subjects. While not sufficiently advanced for use as a replacement to spectroscopy the …


Software Updates To A Multiple Autonomous Quadcopter Search System (Maqss), Jared Speck, Toby Chan May 2017

Software Updates To A Multiple Autonomous Quadcopter Search System (Maqss), Jared Speck, Toby Chan

Computer Engineering

A series of performance-based and feature implementation software updates to an existing multiple vehicle autonomous target search system is outlined in this paper. The search system, MAQSS, is designed to address a computational power constraint found on modern autonomous aerial platforms by separating real-time and computationally expensive tasks through delegation to multiple multirotor vehicles. A Ground Control Station (GCS) is also described as part of the MAQSS system to perform the delegation and provide a low workload user interface. Ultimately, the changes to MAQSS noted in this paper helped to improve the performance of the autonomous search mission, the accuracy …


Using Intel Realsense Depth Data For Hand Tracking In Unreal Engine 4, Granger Lang Mar 2017

Using Intel Realsense Depth Data For Hand Tracking In Unreal Engine 4, Granger Lang

Liberal Arts and Engineering Studies

This project describes how to build a hand tracking method for VR/AR using the raw data from a depth sensing camera.


Development Of A Tridimensional Measuring Application For Ipads, Michael Casebolt, Nicolas Kouatli, Jack Mullen May 2015

Development Of A Tridimensional Measuring Application For Ipads, Michael Casebolt, Nicolas Kouatli, Jack Mullen

Computer Science and Software Engineering

In today’s fast-paced distribution centers workers and management alike are constantly searching for the quickest and most efficient way to package items for distribution. Even with the advancement of app-oriented solutions to a variety of problems across many industries there is a distinct unmet need in distribution environments for an application capable of increasing the efficiency and accuracy of packaging items. This senior project focused on the development and testing of an application utilizing the Structure Three Dimensional Sensor and a 4th generation iPad to scan an object or group of objects to be packaged and determine the overall dimensions …


Early Forest Fire Heat Plume Detection Using Neural Network Classification Of Spectral Differences Between Long-Wave And Mid-Wave Infrared Regions, Raul-Alexander Aldama Jun 2013

Early Forest Fire Heat Plume Detection Using Neural Network Classification Of Spectral Differences Between Long-Wave And Mid-Wave Infrared Regions, Raul-Alexander Aldama

Master's Theses

It is difficult to capture the early signs of a forest fire at night using current visible-spectrum sensor technology. Infrared (IR) light sensors, on the other hand, can detect heat plumes expelled at the initial stages of a forest fire around the clock. Long-wave IR (LWIR) is commonly referred to as the “thermal infrared” region where thermal emissions are captured without the need of, or reflections from, external radiation sources. Mid‑wave IR (MWIR) bands lie between the “thermal infrared” and “reflected infrared” (i.e. short-wave IR) regions. Both LWIR and MWIR spectral regions are able to detect thermal radiation; however, they …


California Polytechnic State University Senior Project Winter-Spring 2012, Roborodentia Xvii, Stack-E, Alejandro Ignacio, Austin Hobbs Jun 2012

California Polytechnic State University Senior Project Winter-Spring 2012, Roborodentia Xvii, Stack-E, Alejandro Ignacio, Austin Hobbs

Computer Engineering

The main goal for our project is to design and build a functional autonomous robot that is capable of navigating an open arena while avoiding obstacles, as well as identify other objects or cans on the field. It must also be capable of stacking and containing these cans. Deliverables will include the fully assembled robot chassis containing the essential hardware components needed to accomplish the navigation and movement, as well as capabilities like identification of objects and stacking of cans. Alongside the hardware, there will also be software developed to showcase these capabilities of the robot design, including the vision …


Markerless Affine Region Tracking And Augmentation Using Mser And Sift, Gregory (Greg) Eddington Ii Jun 2011

Markerless Affine Region Tracking And Augmentation Using Mser And Sift, Gregory (Greg) Eddington Ii

Computer Engineering

Due to the advancements in mobile computing hardware and the inclusion of cameras in many computing platforms, augmented reality systems have become widely available. This paper presents a real-time implementation of a novel markerless augmented reality algorithm which is able to track two-dimensional affine regions without a priori information of the environment or computing a world model. The implementation consists of the MAR library; a modular software library which performs the region detection, identification, and tracking; and the Lighthouse application; a program which uses the MAR library to allow the user to augment scenes viewed from a camera. The algorithm …