Open Access. Powered by Scholars. Published by Universities.®

Computer Engineering Commons

Open Access. Powered by Scholars. Published by Universities.®

Data Storage Systems

PDF

2015

Institution
Keyword
Publication
Publication Type

Articles 31 - 55 of 55

Full-Text Articles in Computer Engineering

What’S New Since The April 2013 Stim Ir Subcommittee Report To Cold: Hydra, Islandora And Dspace, Aaron Collier, Suzanna Conrad, Carmen Mitchell, Joan Parker, Andrew Weiss, Jeremy C. Shellhase Apr 2015

What’S New Since The April 2013 Stim Ir Subcommittee Report To Cold: Hydra, Islandora And Dspace, Aaron Collier, Suzanna Conrad, Carmen Mitchell, Joan Parker, Andrew Weiss, Jeremy C. Shellhase

Inaugural CSU IR Conference, 2015

Aaron Collier, Digital Repository Services Manager, Chancellor’s Office
Suzanna Conrad, Digital Initiatives Librarian, Cal Poly Pomona
Carmen Mitchell, Institutional Repository Librarian, CSU San Marcos
Joan Parker, Librarian, Moss Landing Marine Laboratories
Andrew Weiss, Digital Services Librarian, CSU Northridge

Jeremy Shellhase, Head of Information Services & Systems Department, Humboldt State University


The State Of Scholarworks, Aaron Collier Apr 2015

The State Of Scholarworks, Aaron Collier

Inaugural CSU IR Conference, 2015

Aaron Collier, Digital Repository Services Manager, Chancellor’s Office


Cumulonimbus Computing Concerns: Information Security In Public, Private, And Hybrid Cloud Computing, Daniel Adams Apr 2015

Cumulonimbus Computing Concerns: Information Security In Public, Private, And Hybrid Cloud Computing, Daniel Adams

Senior Honors Theses

Companies of all sizes operating in all markets are moving toward cloud computing for greater flexibility, efficiency, and cost savings. The decision of how to adopt the cloud is a question of major security concern due to the fact that control is relinquished over certain portions of the IT ecosystem. This thesis presents the position that the main security decision in moving to cloud computing is choosing which type of cloud to employ for each portion of the network – the hybrid cloud approach. Vulnerabilities that exist on a public cloud will be explored, and recommendations on decision factors will …


Optimizing Cloud Use Under Interval Uncertainty, Vladik Kreinovich, Esthela Gallardo Apr 2015

Optimizing Cloud Use Under Interval Uncertainty, Vladik Kreinovich, Esthela Gallardo

Departmental Technical Reports (CS)

One of the main advantages of cloud computing is that it helps the users to save money: instead of buying a lot of computers to cover all their computations, the user can rent the computation time on the cloud to cover the rare peak spikes of computer need. From this viewpoint, it is important to find the optimal division between in-house and in-the-cloud computations. In this paper, we solve this optimization problem, both in the idealized case when we know the complete information about the costs and the user's need, and in a more realistic situation, when we only know …


Energy Agile Cluster Communication, Muhammad Zain Mustafa Mar 2015

Energy Agile Cluster Communication, Muhammad Zain Mustafa

Masters Theses

Computing researchers have long focused on improving energy-efficiency?the amount of computation per joule? under the implicit assumption that all energy is created equal. Energy however is not created equal: its cost and carbon footprint fluctuates over time due to a variety of factors. These fluctuations are expected to in- tensify as renewable penetration increases. Thus in my work I introduce energy-agility a design concept for a platform?s ability to rapidly and efficiently adapt to such power fluctuations. I then introduce a representative application to assess energy-agility for the type of long-running, parallel, data-intensive tasks that are both common in data …


Universal Schema For Knowledge Representation From Text And Structured Data, Limin Yao Mar 2015

Universal Schema For Knowledge Representation From Text And Structured Data, Limin Yao

Doctoral Dissertations

In data integration we transform information from a source into a target schema. A general problem in this task is loss of fidelity and coverage: the source expresses more knowledge than that can be fit into the target schema, or knowledge that is hard to fit into any schema at all. This problem is taken to an extreme in information extraction (IE) where the source is natural language---one of the most expressive forms of knowledge representation. To address this issue, one can either automatically learn a latent schema emergent in text (a brittle and ill-defined task), or manually define schemas. …


Cybersecurity In The Information Age Mar 2015

Cybersecurity In The Information Age

DePaul Magazine

According to the Identify Theft Resource Center, more than 750 breaches occurred in 2014, exposing more than 83 million records. Breaches occur because people have discovered that there’s money to be made from stealing corporate and customer information. Faced with this reality, consumers and companies may wonder if it’s possible to protect their personal and financial data. DePaul's College of Computing and Digital Media is training the next generation of cybersecurity experts to thwart cybersecurity attacks. In addition, DePaul faculty and alumni in data security offer advice to consumers on securing their financial and personal information.


Click-Boosting Multi-Modality Graph-Based Reranking For Image Search, Xiaopeng Yang, Yongdong Zhang, Ting Yao, Chong-Wah Ngo, Tao Mei Mar 2015

Click-Boosting Multi-Modality Graph-Based Reranking For Image Search, Xiaopeng Yang, Yongdong Zhang, Ting Yao, Chong-Wah Ngo, Tao Mei

Research Collection School Of Computing and Information Systems

Image reranking is an effective way for improving the retrieval performance of keyword-based image search engines. A fundamental issue underlying the success of existing image reranking approaches is the ability in identifying potentially useful recurrent patterns from the initial search results. Ideally, these patterns can be leveraged to upgrade the ranks of visually similar images, which are also likely to be relevant. The challenge, nevertheless, originates from the fact that keyword-based queries are used to be ambiguous, resulting in difficulty in predicting the search intention. Mining useful patterns without understanding query is risky, and may lead to incorrect judgment in …


Csr: Small: Collaborative Research: Sane: Semantic-Aware Namespace In Exascale File Systems, Yifeng Zhu Feb 2015

Csr: Small: Collaborative Research: Sane: Semantic-Aware Namespace In Exascale File Systems, Yifeng Zhu

University of Maine Office of Research Administration: Grant Reports

Explosive growth in volume and complexity of data exacerbates the key challenge facing the management of massive data in a way that fundamentally improves the ease and efficacy of their usage. Exascale storage systems in general rely on hierarchically structured namespace that leads to severe performance bottlenecks and makes it hard to support real-time queries on multi-dimensional attributes. Thus, existing storage systems, characterized by the hierarchical directory tree structure, are not scalable in light of the explosive growth in both the volume and the complexity of data. As a result, directory-tree based hierarchical namespace has become restrictive, difficult to use, …


Contextual Anomaly Detection Framework For Big Sensor Data, Michael Hayes, Miriam A M Capretz Jan 2015

Contextual Anomaly Detection Framework For Big Sensor Data, Michael Hayes, Miriam A M Capretz

Miriam A M Capretz

The ability to detect and process anomalies for Big Data in real-time is a difficult task. The volume and velocity of the data within many systems makes it difficult for typical algorithms to scale and retain their real-time characteristics. The pervasiveness of data combined with the problem that many existing algorithms only consider the content of the data source; e.g. a sensor reading itself without concern for its context, leaves room for potential improvement. The proposed work defines a contextual anomaly detection framework. It is composed of two distinct steps: content detection and context detection. The content detector is used …


Hadoop Based Data Intensive Computation On Iaas Cloud Platforms, Sruthi Vijayakumar Jan 2015

Hadoop Based Data Intensive Computation On Iaas Cloud Platforms, Sruthi Vijayakumar

UNF Graduate Theses and Dissertations

Cloud computing is a relatively new form of computing which uses virtualized resources. It is dynamically scalable and is often provided as pay for use service over the Internet or Intranet or both. With increasing demand for data storage in the cloud, the study of data-intensive applications is becoming a primary focus. Data intensive applications are those which involve high CPU usage, processing large volumes of data typically in size of hundreds of gigabytes, terabytes or petabytes. The research in this thesis is focused on the Amazon’s Elastic Cloud Compute (EC2) and Amazon Elastic Map Reduce (EMR) using HiBench Hadoop …


Sheep Updates 2015 - Merredin, Bruce Mullan, Kate Pritchett, Kimbal Curtis, Chris Wilcox, Lynne Bradshaw, Geoff Lindon, Katherine Davies, Joe Young, Stephen Lee, Dawson Bradford, Khama Kelman, Lucy Anderton, Jaq Pearson, Jackie Jarvis, Ben Patrick Jan 2015

Sheep Updates 2015 - Merredin, Bruce Mullan, Kate Pritchett, Kimbal Curtis, Chris Wilcox, Lynne Bradshaw, Geoff Lindon, Katherine Davies, Joe Young, Stephen Lee, Dawson Bradford, Khama Kelman, Lucy Anderton, Jaq Pearson, Jackie Jarvis, Ben Patrick

Sheep Updates

This session covers fourteen papers from different authors:

1. The Sheep Industry Business Innovation project, Bruce Mullan, Sheep Industry Development Director, Department of Agriculture and Food, Western Australia

2. Western Australian sheep stocktake, Kate Pritchett and Kimbal Curtis, Research Officers, Department of Agriculture and Food, Western Australia

3. Wool demand and supply - short term volatility, long term opportunities, Chris Wilcox, Principal of Poimena Analysis

4. Myths, Facts and the role of animal welfare in farming, Lynne Bradshaw, president, RSPCA WA

5. Latest research and development on breech strike prevention, Geoff Lindon, Manager Productivity and Animal Welfare, AWI

6. …


Table Of Contents Jan 2015

Table Of Contents

Journal of International Technology and Information Management

No abstract provided.


From The Desk Of The Editor Jan 2015

From The Desk Of The Editor

Journal of International Technology and Information Management

No abstract provided.


Table Of Contents Jan 2015

Table Of Contents

Journal of International Technology and Information Management

No abstract provided.


From The Desk Of The Editor Jan 2015

From The Desk Of The Editor

Journal of International Technology and Information Management

No abstract provided.


Table Of Contents Jan 2015

Table Of Contents

Journal of International Technology and Information Management

No abstract provided.


From The Desk Of The Editor Jan 2015

From The Desk Of The Editor

Journal of International Technology and Information Management

No abstract provided.


Table Of Contents Jan 2015

Table Of Contents

Journal of International Technology and Information Management

No abstract provided.


From The Desk Of The Editor Jan 2015

From The Desk Of The Editor

Journal of International Technology and Information Management

No abstract provided.


Evolution And Usage Of The Portal Data Archive: 10-Year Retrospective, Kristin A. Tufte, Robert Bertini, Morgan Harvey Jan 2015

Evolution And Usage Of The Portal Data Archive: 10-Year Retrospective, Kristin A. Tufte, Robert Bertini, Morgan Harvey

Civil and Environmental Engineering Faculty Publications and Presentations

The Portal transportation data archive (http://portal.its.pdx.edu/) was begun in June 2004 in collaboration with the Oregon Department of Transportation, with a single data source: freeway loop detector data. In 10 years, Portal has grown to contain approximately 3 TB of transportation-related data from a wide variety of systems and sources, including freeway data, arterial signal data, travel times from Bluetooth detection systems, transit data, and bicycle count data. Over its 10-year existence, Portal has expanded both in the type of data that it receives and in the geographic regions from which it gets data. This paper discusses the …


Automated Beverage Dispenser, Sonya Istocka Jan 2015

Automated Beverage Dispenser, Sonya Istocka

Williams Honors College, Honors Research Projects

The intention of this project is to define a new way of distributing liquor. The project will consist of a device which will measure and track liquor being poured and associate it with a person, either a bartender or a bar patron. The challenges will be controlling the flow of liquor and recording it in an extremely accurate manner as well as processing data quickly so that a pour can be initiated very soon after a person is identified. The liquor dispenser will open up the possibility of a person being able to dispense their own liquor in a controlled …


Comparing The Efficiency Of Heterogeneous And Homogeneous Data Center Workloads, Brandon Kimmons Jan 2015

Comparing The Efficiency Of Heterogeneous And Homogeneous Data Center Workloads, Brandon Kimmons

Electronic Theses and Dissertations

Abstract

Information Technology, as an industry, is growing very quickly to keep pace with increased data storage and computing needs. Data growth, if not planned or managed correctly, can have larger efficiency implications on your data center as a whole. The long term reduction in efficiency will increase costs over time and increase operational overhead. Similarly, increases in processor efficiency have led to increased system density in data centers. This can increase cost and operational overhead in your data center infrastructure.

This paper proposes the idea that balanced data center workloads are more efficient in comparison to similar levels of …


Testing Data Vault-Based Data Warehouse, Connard N. Williams Jan 2015

Testing Data Vault-Based Data Warehouse, Connard N. Williams

Electronic Theses and Dissertations

Data warehouse (DW) projects are undertakings that require integration of disparate sources of data, a well-defined mapping of the source data to the reconciled data, and effective Extract, Transform, and Load (ETL) processes. Owing to the complexity of data warehouse projects, great emphasis must be placed on an agile-based approach with properly developed and executed test plans throughout the various stages of designing, developing, and implementing the data warehouse to mitigate against budget overruns, missed deadlines, low customer satisfaction, and outright project failures. Yet, there are often attempts to test the data warehouse exactly like traditional back-end databases and legacy …


Hair-Oriented Data Model For Spatio-Temporal Data Mining, Abbas Madraky, Zulaiha Ali Othman, Razak Hamdan Dec 2014

Hair-Oriented Data Model For Spatio-Temporal Data Mining, Abbas Madraky, Zulaiha Ali Othman, Razak Hamdan

Abbas Madraky

Spatio-temporal data are complex in terms of number of attributes for spatial and temporal values, and the data are changing towards time. Traditional method to mining the spatio-temporal data is the fact that the data is stored in data warehouse in un-normalization form as union of spatial and temporal data know as tabular data warehouse. A Hair-Oriented Data Model (HODM) has been proved as a suitable data model for spatio-temporal data. It has reduced the file size and decreased query execution time. The spatio-temporal data stored using the HODM known as Hair-Oriented Data warehouse. However, this paper aims to presents …