Open Access. Powered by Scholars. Published by Universities.®

Computer Sciences Commons

Open Access. Powered by Scholars. Published by Universities.®

31,458 Full-Text Articles 31,455 Authors 7,711,165 Downloads 298 Institutions

All Articles in Computer Sciences

Faceted Search

31,458 full-text articles. Page 860 of 908.

Automatically Incorporating New Sources In Keyword Search-Based Data Integration, Partha Talukdar, Zachary G. Ives, Fernando Pereira 2010 University of Pennsylvania

Automatically Incorporating New Sources In Keyword Search-Based Data Integration, Partha Talukdar, Zachary G. Ives, Fernando Pereira

Departmental Papers (CIS)

Scientific data offers some of the most interesting challenges in data integration today. Scientific fields evolve rapidly and accumulate masses of observational and experimental data that needs to be annotated, revised, interlinked, and made available to other scientists. From the perspective of the user, this can be a major headache as the data they seek may initially be spread across many databases in need of integration. Worse, even if users are given a solution that integrates the current state of the source databases, new data sources appear with new data items of interest to the user. Here we build upon ...


Automatic Junit Creation Tool: An Exploration In High Level Process Driven Automatic Test Case Creation, William Whitney 2010 California Polytechnic State University - San Luis Obispo

Automatic Junit Creation Tool: An Exploration In High Level Process Driven Automatic Test Case Creation, William Whitney

Computer Science

Many software developers do not enjoy writing unit test code. Often their excuses range from testing is slow to testing is hard. Yet perhaps test derivation has to be neither. The aim of this senior project is to examine the current state of unit test creation for the Java programming language. In particular, inefficiencies with the JUnit test framework regarding test derivation are analyzed. Ultimately, a JUnit test creation tool is created that provides a high-level process for test derivation.


Optimization Of P3ht-Pcbm Polymer Solar Cells Through Device Simulation And Manufacturing, James Boom 2010 California Polytechnic State University - San Luis Obispo

Optimization Of P3ht-Pcbm Polymer Solar Cells Through Device Simulation And Manufacturing, James Boom

Computer Engineering

Given a good model and implementation of that model, computer simulation can be used to reduce the time and material costs of research. To this end I worked with other students to manufacture, test and simulate the single layer P3HT-PCBM solar cell. Using the data collected from this project, future work can then be done with the project's simulator to further optimize these types of solar cell devices.


Max Flow Spill Code Placement Algorithm Implemented In Gcc 4.4.3, Stephen Robert Beard 2010 CaliforniaPolytechnic State University - San Luis Obispo

Max Flow Spill Code Placement Algorithm Implemented In Gcc 4.4.3, Stephen Robert Beard

Computer Engineering

The placement of spill code plays an important role in the register allocator of an optimizing compiler. Many computer architectures possess a register linkage convention that dictates which registers are preserved across function calls and which are not. This project addresses the problem of optimizing spill code that is associated with register linkage conventions.

This algorithm was created by Dr. Chris Lupo and is described in the paper Beyond Register Allocation: a Novel Algorithm for Spill-Code Placement. The algorithm was implemented for GCC 2.5.7 for a PA-RISC architecture [4]. The work in this project will involve porting the ...


Measurement And Interpolation Of Sea Surface Temperature And Salinity In The Tropical Pacific: A 9,000 Nautical Mile Research Odyssey, Amber Brooks 2010 California Polytechnic State University - San Luis Obispo

Measurement And Interpolation Of Sea Surface Temperature And Salinity In The Tropical Pacific: A 9,000 Nautical Mile Research Odyssey, Amber Brooks

Earth and Soil Sciences

The purpose of this project was to compare spline and inverse distance weighting interpolation tools on data collected in the tropical Pacific Ocean by ship and data from a global network of CTD floats, known as Argo floats (fig.1), to provide evidence that technological advancement and integration is aiding our understanding of the ocean-atmosphere system of planet Earth. Thirty-one sea surface temperature and salinity samples were manually taken across a 9,000 nautical mile trek of the Pacific Ocean for the months of April, May and June 2008. Argo ASCII globally gridded monthly averaged sea surface temperature and salinity ...


Wii-Mote Head Tracking: A Three Dimensional Virtual Reality Display, David Fairman 2010 California Polytechnic State University - San Luis Obispo

Wii-Mote Head Tracking: A Three Dimensional Virtual Reality Display, David Fairman

Computer Engineering

The goal of this project is to create a customizable three dimensional virtual reality display on a system available to any non-technical user. This System will use the infrared camera component of a standard Nintendo Wii-mote to track a user's head motions in all six major directions. The virtual reality will be a customizable image projected onto a screen or simply shown on a computer or TV monitor. In order to appear 3-dimensional, the image will continually change according to the position of the user's head. As the user moves their head to the left and right, portions ...


Real-Time Visualizations Of Ocean Data Collected By The Norus Glider, Daniel M. Medina 2010 California Polytechnic State University, San Luis Obispo

Real-Time Visualizations Of Ocean Data Collected By The Norus Glider, Daniel M. Medina

Master's Theses and Project Reports

Scientific visualization computer applications generate visual representations of large and complex sets of science data. These types of applications allow scientists to gain greater knowledge and insight into their data. For example, the visualization of environmental data is of particular interest to biologists when trying to understand how complex variables interact. Modern robotics and sensors have expanded the ability to collect environmental data, thus, the size and variety of these data-sets have likewise grown. Oftentimes, the collected data are deposited into files and databases where they sit in their separate and unique formats. Without easy to use visualization tools, it ...


Simulating An Airborne Lidar Bathymetry (Alb) System, Shachak Pe'eri, Amaresh M. Kumar, Brian R. Calder 2010 University of New Hampshire, Durham

Simulating An Airborne Lidar Bathymetry (Alb) System, Shachak Pe'eri, Amaresh M. Kumar, Brian R. Calder

Center for Coastal and Ocean Mapping

This study’s focus is on the horizontal and vertical uncertainties associated with ALB measurements due to scattering through the water column. A lidar simulator was constructed and we present its design and preliminary results.


Information Risk And Underwriter Switching In Seos: Evidence From China, Wei LUO, Pingui RAO, Heng YUE 2010 Singapore Management University

Information Risk And Underwriter Switching In Seos: Evidence From China, Wei Luo, Pingui Rao, Heng Yue

Research Collection School Of Accountancy

In this paper we examine whether information risk affects underwriter switching in a seasoned equity offering (SEO) process. Building on previous research, we hypothesize that SEO firms and underwriters associate with one another by mutual choice, and firms with a low degree of information risk tend to match up with prestigious underwriters. Using a sample of SEO firms in China and employing accruals quality as a proxy of information risk, we find evidence consistent with our hypothesis: the information risk and the initial public offering (IPO) underwriters’ reputation at the time of the SEO jointly determine the probability that the ...


Provenance Management In Parasite Research, Vinh Nguyen, Priti Parikh, Satya S. Sahoo, Amit P. Sheth 2010 Wright State University - Main Campus

Provenance Management In Parasite Research, Vinh Nguyen, Priti Parikh, Satya S. Sahoo, Amit P. Sheth

Kno.e.sis Publications

The objective of this research is to create a semantic problem solving environment (PSE) for human parasite Trypanosoma cruzi. As a part of the PSE, we are trying to manage provenance of the experiment data as it is generated. It requires to capture the provenance which is often collected through web forms used by biologists to input the information about experiments they conduct. We have created Parasite Experiment Ontology (PEO) that represents provenance information used in the project. We have modified the back end which processes the data gathered from biologists, generates RDF triples and serializes them into the triple ...


Graduation Outliner, Jason Boyle 2010 California Polytechnic State University - San Luis Obispo

Graduation Outliner, Jason Boyle

Computer Science

Planning a quarterly course schedule is a difficult and laborious process for university students which must be repeated every quarter. Students are expected to determine which courses to take based on a large and complex set of requirement criteria. Most academic scheduling research has focused on faculty availability and desires. This project attempts to address the needs of students for whom no quarterly course schedule planning tool is widely available. A Java Swing-based application is described that allows students to create a graphical quarter-by-quarter visualization of the courses they plan to take. The application also accepts a list of graduation ...


Customer Communicator, Eddie Tavarez 2010 California Polytechnic State University - San Luis Obispo

Customer Communicator, Eddie Tavarez

Computer Science

No abstract provided.


Employee Time Scheduling, Mark Peter Smith 2010 California Polytechnic State University - San Luis Obispo

Employee Time Scheduling, Mark Peter Smith

Computer Science

Small business managers face the common problem of employee time scheduling. There is a solution to this problem in the form of an application called Lemming Scheduler. Lemming Scheduler is a Java based employee time scheduling program. Its features include a desktop based application that stores employee and business information as well as a web interface for employees to view schedules and update availability. The desktop application uses employee and shift information to automatically generate schedules. The generated schedules are viewable by employees outside of work by way of the web interface. Lemming Scheduler provides a light weight interface for ...


Reducing Cluster Power Consumption By Dynamically Suspending Idle Nodes, Brian Michael Oppenheim 2010 California Polytechnic State University - San Luis Obispo

Reducing Cluster Power Consumption By Dynamically Suspending Idle Nodes, Brian Michael Oppenheim

Master's Theses and Project Reports

Close to 1% of the world's electricity is consumed by computer servers. Given that the increased use of electricity raises costs and damages the environment, optimizing the world's computing infrastructure for power consumption is worthwhile. This thesis is one attempt at such an optimization. In particular, I began by building a cluster of 6 Intel Atom based low-power nodes to perform work analogous to data center clusters. Then, I installed a version of Hadoop modified with a novel power management system on the cluster. The power management system uses different algorithms to determine when to turn off idle ...


Analysis And Transformation Of Pipe-Like Web Mashups For End User Programmers, Kathryn T. Stolee 2010 University of Nebraska at Lincoln

Analysis And Transformation Of Pipe-Like Web Mashups For End User Programmers, Kathryn T. Stolee

Computer Science and Engineering: Theses, Dissertations, and Student Research

Mashups are becoming increasingly popular as end users are able to easily access, manipulate, and compose data from several web sources. To support end users, communities are forming around mashup development environments that facilitate sharing code and knowledge. We have observed, however, that end user mashups tend to suffer from several deficiencies, such as inoperable components or references to invalid data sources, and that those deficiencies are often propagated through the rampant reuse in these end user communities.

In this work, we identify and specify ten code smells indicative of deficiencies we observed in a sample of 8,051 pipe-like ...


Z-Sky: An Efficient Skyline Query Processing Framework Based On Z-Order, Ken C. K. LEE, Wang-chien LEE, Baihua ZHENG, Huajing LI, Yuan TIAN 2010 Pennsylvania State University

Z-Sky: An Efficient Skyline Query Processing Framework Based On Z-Order, Ken C. K. Lee, Wang-Chien Lee, Baihua Zheng, Huajing Li, Yuan Tian

Research Collection School Of Information Systems

Given a set of data points in a multidimensional space, a skyline query retrieves those data points that are not dominated by any other point in the same dataset. Observing that the properties of Z-order space filling curves (or Z-order curves) perfectly match with the dominance relationships among data points in a geometrical data space, we, in this paper, develop and present a novel and efficient processing framework to evaluate skyline queries and their variants, and to support skyline result updates based on Z-order curves. This framework consists of ZBtree, i.e., an index structure to organize a source dataset ...


Weakly-Supervised Hashing In Kernel Space, Yadong MU, Jialie SHEN, Shuicheng YAN 2010 National University of Singapore

Weakly-Supervised Hashing In Kernel Space, Yadong Mu, Jialie Shen, Shuicheng Yan

Research Collection School Of Information Systems

The explosive growth of the vision data motivates the recent studies on efficient data indexing methods such as locality-sensitive hashing (LSH). Most existing approaches perform hashing in an unsupervised way. In this paper we move one step forward and propose a supervised hashing method, i.e., the LAbel-regularized Max-margin Partition (LAMP) algorithm. The proposed method generates hash functions in weakly-supervised setting, where a small portion of sample pairs are manually labeled to be “similar” or “dissimilar”. We formulate the task as a Constrained Convex-Concave Procedure (CCCP), which can be relaxed into a series of convex sub-problems solvable with efficient Quadratic-Program ...


Player Performance Prediction In Massively Multiplayer Online Role-Playing Games (Mmorpgs), Kyong Jin SHIM, R. SHARAN, J. SRIVASTAVA 2010 Singapore Management University

Player Performance Prediction In Massively Multiplayer Online Role-Playing Games (Mmorpgs), Kyong Jin Shim, R. Sharan, J. Srivastava

Research Collection School Of Information Systems

In this study, we propose a comprehensive performance management tool for measuring and reporting operational activities of game players. This study uses performance data of game players in EverQuest II, a popular MMORPG developed by Sony Online Entertainment, to build performance prediction models forgame players. The prediction models provide a projection of player’s future performance based on his past performance, which is expected to be a useful addition to existing player performance monitoring tools. First, we show that variations of PECOTA [2] and MARCEL [3], two most popular baseball home run prediction methods, can be used for game player ...


On Trustworthiness Of Cpu Usage Metering And Accounting, Mei LIU, Xuhua DING 2010 Singapore Management University

On Trustworthiness Of Cpu Usage Metering And Accounting, Mei Liu, Xuhua Ding

Research Collection School Of Information Systems

In the envisaged utility computing paradigm, a user taps a service provider’s computing resources to accomplish her tasks, without deploying the needed hardware and software in her own IT infrastructure. To make the service profitable, the service provider charges the user based on the resources consumed. A commonly billed resource is CPU usage. A key factor to ensure the success of such a business model is the trustworthiness of the resource metering scheme. In this paper, we provide a systematic study on the trustworthiness of CPU usage metering. Our results show that the metering schemes in commodity operating systems ...


Prediction Of Protein Subcellular Localization: A Machine Learning Approach, Kyong Jin SHIM 2010 Singapore Management University

Prediction Of Protein Subcellular Localization: A Machine Learning Approach, Kyong Jin Shim

Research Collection School Of Information Systems

Subcellular localization is a key functional characteristic of proteins. Optimally combining available information is one of the key challenges in today's knowledge-based subcellular localization prediction approaches. This study explores machine learning approaches for the prediction of protein subcellular localization that use resources concerning Gene Ontology and secondary structures. Using the spectrum kernel for feature representation of amino acid sequences and secondary structures, we explore an SVM-based learning method that classifies six subcellular localization sites: endoplasmic reticulum, extracellular, Golgi, membrane, mitochondria, and nucleus.


Digital Commons powered by bepress