Open Access. Powered by Scholars. Published by Universities.®

Digital Commons Network

Open Access. Powered by Scholars. Published by Universities.®

Engineering

PDF

Metadata

Institution
Publication Year
Publication
Publication Type

Articles 1 - 30 of 34

Full-Text Articles in Entire DC Network

Enhancing Mobile App User Experience: A Deep Learning Approach For System Design And Optimization, Deepesh Haryani Apr 2024

Enhancing Mobile App User Experience: A Deep Learning Approach For System Design And Optimization, Deepesh Haryani

Harrisburg University Dissertations and Theses

This paper presents a comprehensive framework for enhancing user experience in mobile applications through the integration of deep learning systems. The proposed system design encompasses various components, including data collection and preprocessing, model development and training, integration with mobile applications, dataset management service, model training service, model serving, hyperparameter optimization, metadata and artifact store, and workflow orchestration. Each component is meticulously designed with a focus on scalability, efficiency, isolation, and critical analysis. Innovative design principles are employed to ensure seamless integration, usability, and automation. Additionally, the paper discusses distributed training service design, advanced optimization techniques, and decision criteria for hyperparameter …


Hashes Are Not Suitable To Verify Fixity Of The Public Archived Web, Mohamed Aturban, Martin Klein, Herbert Van De Sompel, Sawood Alam, Michael L. Nelson, Michele C. Weigle Jan 2023

Hashes Are Not Suitable To Verify Fixity Of The Public Archived Web, Mohamed Aturban, Martin Klein, Herbert Van De Sompel, Sawood Alam, Michael L. Nelson, Michele C. Weigle

Computer Science Faculty Publications

Web archives, such as the Internet Archive, preserve the web and allow access to prior states of web pages. We implicitly trust their versions of archived pages, but as their role moves from preserving curios of the past to facilitating present day adjudication, we are concerned with verifying the fixity of archived web pages, or mementos, to ensure they have always remained unaltered. A widely used technique in digital preservation to verify the fixity of an archived resource is to periodically compute a cryptographic hash value on a resource and then compare it with a previous hash value. If the …


Accessing 3d Data, Francesca Albrezzi, John Bonnett, Tassie Gniady, Heather Richards-Rissetto, Lisa Snyder Jan 2022

Accessing 3d Data, Francesca Albrezzi, John Bonnett, Tassie Gniady, Heather Richards-Rissetto, Lisa Snyder

Department of Anthropology: Faculty Publications

The issue of access and discoverability is not simply a matter of permissions and availability. To identify, locate, retrieve, and reuse 3D materials requires consideration of a multiplicity of content types, as well as community and financial investment to resolve challenges related to usability, interoperability, sustainability, and equity. This chapter will cover modes, audiences, assets and decision points, technology requirements, and limitations impacting access, as well as providing recommendations for next steps.


Deep Understanding Of Technical Documents : Automated Generation Of Pseudocode From Digital Diagrams & Analysis/Synthesis Of Mathematical Formulas, Nikolaos Gkorgkolis Jan 2022

Deep Understanding Of Technical Documents : Automated Generation Of Pseudocode From Digital Diagrams & Analysis/Synthesis Of Mathematical Formulas, Nikolaos Gkorgkolis

Browse all Theses and Dissertations

The technical document is an entity that consists of several essential and interconnected parts, often referred to as modalities. Despite the extensive attention that certain parts have already received, per say the textual information, there are several aspects that severely under researched. Two such modalities are the utility of diagram images and the deep automated understanding of mathematical formulas. Inspired by existing holistic approaches to the deep understanding of technical documents, we develop a novel formal scheme for the modelling of digital diagram images. This extends to a generative framework that allows for the creation of artificial images and their …


Fedoram: A Federated Oblivious Ram Scheme, Alexandre Pujol, Liam Murphy, Christina Thorpe Jan 2021

Fedoram: A Federated Oblivious Ram Scheme, Alexandre Pujol, Liam Murphy, Christina Thorpe

Articles

Instant messaging (IM) applications, even with end-to-end encryption enabled, pose privacy issues due to metadata and pattern leakage. Our goal is to develop a model for a privacy preserving IM application, by designing an IM application that focuses on hiding metadata and discussion patterns. To solve the issue of privacy preservation through the obfuscation of metadata, cryptographic constructions like Oblivious Random Access Machines (ORAM) have been proposed in recent years. However, although they completely hide the user access patterns, they incur high computational costs, often resulting in excessively slow performance in practice. We propose a new federated model, FedORAM, which …


Conference Roundup: Smart Cataloging - Beginning The Move From Batch Processing To Automated Classification, Rachel S. Evans Jun 2020

Conference Roundup: Smart Cataloging - Beginning The Move From Batch Processing To Automated Classification, Rachel S. Evans

Articles, Chapters and Online Publications

This article reviewed the Amigos Online Conference titled “Work Smarter, Not Harder: Innovating Technical Services Workflows” keynote session delivered by Dr. Terry Reese on February 13, 2020. Excerpt:

"As the developer of MarcEdit, a popular metadata suite used widely across the library community, Reese’s current work is focused on the ways in which libraries might leverage semantic web techniques in order to transform legacy library metadata into something new. So many sessions related to using new technologies in libraries or academia, although exciting, are not practical enough to put into everyday use by most librarians. Reese’s keynote, titled Smart Cataloging: …


Metadata Management For Clinical Data Integration, Ningzhou Zeng Jan 2020

Metadata Management For Clinical Data Integration, Ningzhou Zeng

Theses and Dissertations--Computer Science

Clinical data have been continuously collected and growing with the wide adoption of electronic health records (EHR). Clinical data have provided the foundation to facilitate state-of-art researches such as artificial intelligence in medicine. At the same time, it has become a challenge to integrate, access, and explore study-level patient data from large volumes of data from heterogeneous databases. Effective, fine-grained, cross-cohort data exploration, and semantically enabled approaches and systems are needed. To build semantically enabled systems, we need to leverage existing terminology systems and ontologies. Numerous ontologies have been developed recently and they play an important role in semantically enabled …


Autoplug: An Automated Metadata Service For Smart Outlets, Lurdh Pradeep Reddy Ambati Oct 2017

Autoplug: An Automated Metadata Service For Smart Outlets, Lurdh Pradeep Reddy Ambati

Masters Theses

Low-cost network-connected smart outlets are now available for monitoring, controlling, and scheduling the energy usage of electrical devices. As a result, such smart outlets are being integrated into automated home management systems, which remotely control them by analyzing and interpreting their data. However, to effectively interpret data and control devices, the system must know the type of device that is plugged into each smart outlet. Existing systems require users to manually input and maintain the outlet metadata that associates a device type with a smart outlet. Such manual operation is time-consuming and error-prone: users must initially inventory all outlet-to-device mappings, …


A Novel Privacy Preserving User Identification Approach For Network Traffic, Nathan Clarke, Fudong Li, Steven Furnell Sep 2017

A Novel Privacy Preserving User Identification Approach For Network Traffic, Nathan Clarke, Fudong Li, Steven Furnell

Research outputs 2014 to 2021

The prevalence of the Internet and cloud-based applications, alongside the technological evolution of smartphones, tablets and smartwatches, has resulted in users relying upon network connectivity more than ever before. This results in an increasingly voluminous footprint with respect to the network traffic that is created as a consequence. For network forensic examiners, this traffic represents a vital source of independent evidence in an environment where anti-forensics is increasingly challenging the validity of computer-based forensics. Performing network forensics today largely focuses upon an analysis based upon the Internet Protocol (IP) address – as this is the only characteristic available. More typically, …


Blown To Bits Project, David Schmidt Jan 2017

Blown To Bits Project, David Schmidt

Informatics Open Educational Resources

The book, Blown to Bits, uncovers the many ways that the new digital world has changed and is changing our whole environment. Some changes are incremental but others are more revolutionary. Some of the changes that we welcome are slowly eroding our privacy and are changing the rules of ownership. This book illuminates the complexities of these changes. I have attempted to capture the central points in selected chapters, and in some cases I have added new material or new examples to replace dated material. I picked chapters to summarize that address the following topics (and more). There are many …


Flexible Analysis Of Creep Rupture Database And Accelerating The Acquisition Of Creep Rupture Data, Christopher Ramirez Jan 2017

Flexible Analysis Of Creep Rupture Database And Accelerating The Acquisition Of Creep Rupture Data, Christopher Ramirez

Open Access Theses & Dissertations

A comprehensive statistical analysis of creep data is a difficult task because there are many layers of uncertainty for a given dataset. Sources of uncertainty are inherent in both the databases that provide data, and the data themselves. Additionally, creep rupture predictions made with time-temperature parameter (TTP) models add an additional layer of uncertainty due to the fundamentally different ways in which each TTP model predicts creep behavior. A set of guidelines from the ECCC currently exist for such analyses, but they are best suited for narrowly-defined datasets. In this study, a broader set of guidelines are developed to analyze …


Digital Preservation Efforts At Usm, Elizabeth La Beaud Nov 2016

Digital Preservation Efforts At Usm, Elizabeth La Beaud

Central Plains Network for Digital Asset Management

The University of Southern Mississippi (USM) has steadily been working to improve its digital preservation infrastructure over the past four years. In 2013, with funding from a NEH Preservation Assistance Grant, consultants Tom Clareson and Liz Bishoff conducted a digital preservation readiness assessment and jump started USM’s education on the topic. Since then, USM has added geographically distributed backups, manual fixity checks, manual metadata logs, and manual file format migrations to its arsenal with varying degrees of success. The influx in needed manpower and technical infrastructure precipitated a financial commitment from the university and the purchase of a robust digital …


An Automated Approach For Digital Forensic Analysis Of Heterogeneous Big Data, Hussam Mohammed, Nathan Clarke, Fudong Li Jan 2016

An Automated Approach For Digital Forensic Analysis Of Heterogeneous Big Data, Hussam Mohammed, Nathan Clarke, Fudong Li

Journal of Digital Forensics, Security and Law

The major challenges with big data examination and analysis are volume, complex interdependence across content, and heterogeneity. The examination and analysis phases are considered essential to a digital forensics process. However, traditional techniques for the forensic investigation use one or more forensic tools to examine and analyse each resource. In addition, when multiple resources are included in one case, there is an inability to cross-correlate findings which often leads to inefficiencies in processing and identifying evidence. Furthermore, most current forensics tools cannot cope with large volumes of data. This paper develops a novel framework for digital forensic analysis of heterogeneous …


Making Music More Open: Copyright, Conservation And Access Issues, Marcus R. Wigan Nov 2015

Making Music More Open: Copyright, Conservation And Access Issues, Marcus R. Wigan

Marcus R Wigan

Marcus Wigan, Institute for Social Research, Swinburne University of Technology and Monash University Sir Zelman Cowan School of Music

Making music more open: copyright, conservation and access issues

IP rights around music have come under increasing stress in the digital age. The demands of consumers for cheap, perpetual access to any and all types of music has clashed with copyright owners and artists seeking to preserve or grow existing revenues. But this emphasis on commercial rights in music has led to one area being neglected. That is music and performance of historical or niche value, especially that held by museums …


Practical Guidance For Integrating Data Management Into Long-Term Ecological Monitoring Projects, Robert D. Sutter, Susan Wainscott, John R. Boetsch, Craig Palmer, David J. Rugg Sep 2015

Practical Guidance For Integrating Data Management Into Long-Term Ecological Monitoring Projects, Robert D. Sutter, Susan Wainscott, John R. Boetsch, Craig Palmer, David J. Rugg

Library Faculty Publications

Long-term monitoring and research projects are essential to understand ecological change and the effectiveness of management activities. An inherent characteristic of long-term projects is the need for consistent data collection over time, requiring rigorous attention to data management and quality assurance. Recent papers have provided broad recommendations for data management; however, practitioners need more detailed guidance and examples. We present general yet detailed guidance for the development of comprehensive, concise, and effective data management for monitoring projects. The guidance is presented as a graded approach, matching the scale of data management to the needs of the organization and the complexity …


Hydrologic Data Sharing Using Open Source Software And Low-Cost Electronics, Jeffrey Michael Sadler Mar 2015

Hydrologic Data Sharing Using Open Source Software And Low-Cost Electronics, Jeffrey Michael Sadler

Theses and Dissertations

While it is generally accepted that environmental data are critical to understanding environmental phenomena, there are yet improvements to be made in their consistent collection, curation, and sharing. This thesis describes two research efforts to improve two different aspects of hydrologic data collection and management. First described is a recipe for the design, development, and deployment of a low-cost environmental data logging and transmission system for environmental sensors and its connection to an open source data-sharing network. The hardware is built using several low-cost, open-source, mass-produced components. The system automatically ingests data into HydroServer, a standards-based server in the open …


Improving Processing By Adaption To Conditional Geostatistical Simulation Of Block Compositions, R. Tolosana-Delgado, Ute A. Mueller, K. G. Van Den Boogaart, C. Ward, J. Gutzmer Jan 2015

Improving Processing By Adaption To Conditional Geostatistical Simulation Of Block Compositions, R. Tolosana-Delgado, Ute A. Mueller, K. G. Van Den Boogaart, C. Ward, J. Gutzmer

Research outputs 2014 to 2021

Exploitation of an ore deposit can be optimized by adapting the beneficiation processes to the properties of individual ore blocks. This can involve switching in and out certain treatment steps, or setting their controlling parameters. Optimizing this set of decisions requires the full conditional distribution of all relevant physical parameters and chemical attributes of the feed, including concentration of value elements and abundance of penalty elements. As a first step towards adaptive processing, the mapping of adaptive decisions is explored based on the composition, in value and penalty elements, of the selective mining units. Conditional distributions at block support are …


Agriculture And Biological Engineering/ Graduate Students/ Bracke & Fosmire/ Purdue University/ 2012, Marianne S. Bracke, Michael Fosmire Jan 2015

Agriculture And Biological Engineering/ Graduate Students/ Bracke & Fosmire/ Purdue University/ 2012, Marianne S. Bracke, Michael Fosmire

Data Information Literacy Case Study Directory

This case study was comprised of three workshops for a lab in Agricultural and Biological Engineering. Data Information Literacy topics covered include developing and following a standard operating procedure, search for external data and creating metadata. Materials include a book chapter describing the case study, a data archiving checklist to follow for describing the types of data generated by the lab, a metadata entry form and an evaluation form used to gather feedback for each of the three sessions.


Ontology Mapping In Semantic Time Series Processing In Climate Change Prediction, Bojan Božić, Jan Peters-Anders, Gerald Schimak Jun 2014

Ontology Mapping In Semantic Time Series Processing In Climate Change Prediction, Bojan Božić, Jan Peters-Anders, Gerald Schimak

International Congress on Environmental Modelling and Software

In today's time series processing there is more and more a need for addressing diverse user groups interested in a specific domain with appropriate user tailored time series data. The complexity of time series (e.g. involved data from different data sources and/or domains, visualization and representation, etc.) is growing rapidly. As a consequence, it means that users need to find a path through the jungle of time series data. After we have presented our concepts for semantic time series filtering and enrichment of time series with meta-information and annotations (Božić et al., 2012), we are now going to present a …


Integrating Neon Data With Existing Models: An Example With The Community Land Model, Edmund M. Hart, Andrew Fox, Steve Berukoff, T. J. Hoar Jun 2014

Integrating Neon Data With Existing Models: An Example With The Community Land Model, Edmund M. Hart, Andrew Fox, Steve Berukoff, T. J. Hoar

International Congress on Environmental Modelling and Software

A central challenge to environmental forecasting in hydrological and land surface modeling is how to integrate multiple data sources over a wide range of spatial scales. Furthermore how can this complex task be achieved in the most productive and reproducible way with a robust informatics architecture? At the National Ecological Observatory Network (NEON) we are collecting a variety of biophysical and biogeochemical measurements which can be used with models to perform temporal forecasting on decadal timescales. To take advantage of this data we are developing a data assimilation framework. Using this framework NEON data can be combined with the Community …


Nesll Modular, High Performance Lnfrastructure For Earth System Modeling, Cecelia Deluca, Peggy Li, Gerhard Theurich Jun 2014

Nesll Modular, High Performance Lnfrastructure For Earth System Modeling, Cecelia Deluca, Peggy Li, Gerhard Theurich

International Congress on Environmental Modelling and Software

The mission of the NOAA Environmental Software lnfrastructure and lnteroperability (NESll) group is to advance understanding and improve predictions of the Earth system by delivering infrastructure software that enables new scientific discoveries, fosters collaborations, and promotes resource efficiency. NESll software supports environmental modeling, data search and analysis, and metadata services. A core capability is the Earth System Modeling Framework (ESMF), a comprehensive object-oriented framework for building and coupling models. However, new development has focused on modular packages that can easily be combined with other software elements to form solutions to complex problems. Such modular packages include ESMPy, a Python grid …


Vsfs: A Versatile Searchable File System For Hpc Analytics, Lei Xu, Ziling Huang, Hong Jiang, Lei Tian, David Swanson Apr 2013

Vsfs: A Versatile Searchable File System For Hpc Analytics, Lei Xu, Ziling Huang, Hong Jiang, Lei Tian, David Swanson

CSE Technical Reports

Big-data/HPC analytics applications have urgent needs for file-search services to drastically reduce the scale of the input data to accelerate analytics. Unfortunately, the existing solutions either are poorly scalable for large-scale systems, or lack well-integrated interface to allow applications to easily use them. We propose a distributed searchable file system, VSFS, which provide a novel and flexible POSIX-compatible searchable file system namespace that can be seamlessly integrate with any legacy code without modification. Additionally, to provide real-time indexing and searching performance, VSFS uses DRAM-based distributed consistent hashing ring to manages all file-index. The results of our evaluation show that VSFS …


A Method For Identifying Personalized Representations In Web Archives, Mat Kelly, Justin F. Brunelle, Michele C. Weigle, Michael L. Nelson Jan 2013

A Method For Identifying Personalized Representations In Web Archives, Mat Kelly, Justin F. Brunelle, Michele C. Weigle, Michael L. Nelson

Computer Science Faculty Publications

Web resources are becoming increasingly personalized — two different users clicking on the same link at the same time can see content customized for each individual user. These changes result in multiple representations of a resource that cannot be canonicalized in Web archives. We identify characteristics of this problem by presenting a potential solution to generalize personalized representations in archives. We also present our proof-of-concept prototype that analyzes WARC (Web ARChive) format files, inserts metadata establishing relationships, and provides archive users the ability to navigate on the additional dimension of environment variables in a modified Wayback Machine.


Online Image Tagging System, Malak Alshawabkeh, David Kaeli Apr 2012

Online Image Tagging System, Malak Alshawabkeh, David Kaeli

David Kaeli

Online Image Tagging is a developed software that allows users of the CenSSIS Image Database to associate detailed metadata with specific regions of images, a process we refer to as "Image Tagging." The process of image tagging involves using the Image Tagging interface to draw an outline around a region of interest in an image. Once the area has been selected, a label can be placed on it and additional metadata can be associated with it. The metadata can contain custom XML tags in addition to the standard XML elements that the general tag data is stored in. Eventually, users …


Detecting, Tracking, And Recognizing Activities In Aerial Video, Vladimir Reilly Jan 2012

Detecting, Tracking, And Recognizing Activities In Aerial Video, Vladimir Reilly

Electronic Theses and Dissertations

In this dissertation, we address the problem of detecting humans and vehicles, tracking them in crowded scenes, and finally determining their activities in aerial video. Even though this is a well explored problem in the field of computer vision, many challenges still remain when one is presented with realistic data. These challenges include large camera motion, strong scene parallax, fast object motion, large object density, strong shadows, and insufficiently large action datasets. Therefore, we propose a number of novel methods based on exploiting scene constraints from the imagery itself to aid in the detection and tracking of objects. We show, …


Implementing Transparent Compression And Leveraging Solid State Disks In A High Performance Parallel File System, David Bonnie May 2011

Implementing Transparent Compression And Leveraging Solid State Disks In A High Performance Parallel File System, David Bonnie

All Theses

In recent years computers have been increasing in compute density and speed at a dramatic pace. This increase allows for massively parallel programs to run faster than ever before. Unfortunately, many such programs are being held back by the relatively slow I/O subsystems that they are forced to work with. Storage technology simply has not followed the same curve of progression in the computing world. Because the storage systems are so slow in comparison the processors are forced to idle while waiting for data; a potentially performance crippling condition.
This performance disparity is lessened by the advent of parallel file …


Creating Preservation-Ready Web Resources, Joan A. Smith, Michael L. Nelson Jan 2008

Creating Preservation-Ready Web Resources, Joan A. Smith, Michael L. Nelson

Computer Science Faculty Publications

There are innumerable departmental, community, and personal web sites worthy of long-term preservation but proportionally fewer archivists available to properly prepare and process such sites. We propose a simple model for such everyday web sites which takes advantage of the web server itself to help prepare the site's resources for preservation. This is accomplished by having metadata utilities analyze the resource at the time of dissemination. The web server responds to the archiving repository crawler by sending both the resource and the just-in-time generated metadata as a straight-forward XML-formatted response. We call this complex object (resource + metadata) a CRATE. …


A Jbi Information Object Engineering Environment Utilizing Metadata Fragments For Refining Searches On Semantically-Related Object Types, Felicia N. Harlow Jun 2005

A Jbi Information Object Engineering Environment Utilizing Metadata Fragments For Refining Searches On Semantically-Related Object Types, Felicia N. Harlow

Theses and Dissertations

The Joint Battlespace Infosphere (JBI) architecture defines the Information Object (IO) as its basic unit of data. This research proposes an IO engineering methodology that will introduce componentized IO type development. This enhancement will improve the ability of JBI users to create and store IO type schemas, and query and subscribe to information objects, which may be semantically related by their inclusion of common metadata elements. Several parallel efforts are being explored to enable efficient storage and retrieval of IOs. Utilizing relational database access methods, applying a component-based IO type development concept, and exploiting XML inclusion mechanisms, this research improves …


Lightweight Federation Of Non-Cooperating Digital Libraries, Rong Shi Apr 2005

Lightweight Federation Of Non-Cooperating Digital Libraries, Rong Shi

Computer Science Theses & Dissertations

This dissertation studies the challenges and issues faced in federating heterogeneous digital libraries (DLs). The objective of this research is to demonstrate the feasibility of interoperability among non-cooperating DLs by presenting a lightweight, data driven approach, or Data Centered Interoperability (DCI). We build a Lightweight Federated Digital Library (LFDL) system to provide federated search service for existing digital libraries with no prior coordination.

We describe the motivation, architecture, design and implementation of the LFDL. We develop, deploy, and evaluate key services of the federation. The major difference to existing DL interoperability approaches is one where we do not insist on …


Lessons Learned With Arc, An Oai-Pmh Service Provider, Xiaoming Liu, Kurt Maly, Michael L. Nelson Jan 2005

Lessons Learned With Arc, An Oai-Pmh Service Provider, Xiaoming Liu, Kurt Maly, Michael L. Nelson

Computer Science Faculty Publications

Web-based digital libraries have historically been built in isolation utilizing different technologies, protocols, and metadata. These differences hindered the development of digital library services that enable users to discover information from multiple libraries through a single unified interface. The Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH) is a major, international effort to address technical interoperability among distributed repositories. Arc debuted in 2000 as the first end-user OAI-PMH service provider. Since that time, Arc has grown to include nearly 7,000,000 metadata records. Arc has been deployed in a number of environments and has served as the basis for many other …