Open Access. Powered by Scholars. Published by Universities.®

Computational Linguistics Commons

Open Access. Powered by Scholars. Published by Universities.®

Library and Information Science

Institution
Keyword
Publication Year
Publication
Publication Type

Articles 1 - 12 of 12

Full-Text Articles in Computational Linguistics

Executive Order On The Safe, Secure, And Trustworthy Development And Use Of Artificial Intelligence, Joseph R. Biden Oct 2023

Executive Order On The Safe, Secure, And Trustworthy Development And Use Of Artificial Intelligence, Joseph R. Biden

Copyright, Fair Use, Scholarly Communication, etc.

Section 1. Purpose. Artificial intelligence (AI) holds extraordinary potential for both promise and peril. Responsible AI use has the potential to help solve urgent challenges while making our world more prosperous, productive, innovative, and secure. At the same time, irresponsible use could exacerbate societal harms such as fraud, discrimination, bias, and disinformation; displace and disempower workers; stifle competition; and pose risks to national security. Harnessing AI for good and realizing its myriad benefits requires mitigating its substantial risks. This endeavor demands a society-wide effort that includes government, the private sector, academia, and civil society.

My Administration places the highest urgency …


Chatgpt As Metamorphosis Designer For The Future Of Artificial Intelligence (Ai): A Conceptual Investigation, Amarjit Kumar Singh (Library Assistant), Dr. Pankaj Mathur (Deputy Librarian) Mar 2023

Chatgpt As Metamorphosis Designer For The Future Of Artificial Intelligence (Ai): A Conceptual Investigation, Amarjit Kumar Singh (Library Assistant), Dr. Pankaj Mathur (Deputy Librarian)

Library Philosophy and Practice (e-journal)

Abstract

Purpose: The purpose of this research paper is to explore ChatGPT’s potential as an innovative designer tool for the future development of artificial intelligence. Specifically, this conceptual investigation aims to analyze ChatGPT’s capabilities as a tool for designing and developing near about human intelligent systems for futuristic used and developed in the field of Artificial Intelligence (AI). Also with the helps of this paper, researchers are analyzed the strengths and weaknesses of ChatGPT as a tool, and identify possible areas for improvement in its development and implementation. This investigation focused on the various features and functions of ChatGPT that …


Creating Data From Unstructured Text With Context Rule Assisted Machine Learning (Craml), Stephen Meisenbacher, Peter Norlander Dec 2022

Creating Data From Unstructured Text With Context Rule Assisted Machine Learning (Craml), Stephen Meisenbacher, Peter Norlander

School of Business: Faculty Publications and Other Works

Popular approaches to building data from unstructured text come with limitations, such as scalability, interpretability, replicability, and real-world applicability. These can be overcome with Context Rule Assisted Machine Learning (CRAML), a method and no-code suite of software tools that builds structured, labeled datasets which are accurate and reproducible. CRAML enables domain experts to access uncommon constructs within a document corpus in a low-resource, transparent, and flexible manner. CRAML produces document-level datasets for quantitative research and makes qualitative classification schemes scalable over large volumes of text. We demonstrate that the method is useful for bibliographic analysis, transparent analysis of proprietary data, …


The Public Innovations Explorer: A Geo-Spatial & Linked-Data Visualization Platform For Publicly Funded Innovation Research In The United States, Seth Schimmel Jun 2021

The Public Innovations Explorer: A Geo-Spatial & Linked-Data Visualization Platform For Publicly Funded Innovation Research In The United States, Seth Schimmel

Dissertations, Theses, and Capstone Projects

The Public Innovations Explorer (https://sethsch.github.io/innovations-explorer/app/index.html) is a web-based tool created using Node.js, D3.js and Leaflet.js that can be used for investigating awards made by Federal agencies and departments participating in the Small Business Innovation Research (SBIR) and Small Business Technology Transfer (STTR) grant-making programs between 2008 and 2018. By geocoding the publicly available grants data from SBIR.gov, the Public Innovations Explorer allows users to identify companies performing publicly-funded innovative research in each congressional district and obtain dynamic district-level summaries of funding activity by agency and year. Applying spatial clustering techniques on districts' employment levels across major economic sectors provides users …


Otrouha: A Corpus Of Arabic Etds And A Framework For Automatic Subject Classification, Eman Abdelrahman, Fatimah Alotaibi, Edward A. Fox, Osman Balci Mar 2021

Otrouha: A Corpus Of Arabic Etds And A Framework For Automatic Subject Classification, Eman Abdelrahman, Fatimah Alotaibi, Edward A. Fox, Osman Balci

The Journal of Electronic Theses and Dissertations

Although the Arabic language is spoken by more than 300 million people and is one of the six official languages of the United Nations (UN), there has been less research done on Arabic text data (compared to English) in the realm of machine learning, especially in text classification. In the past decade, Arabic data such as news, tweets, etc. have begun to receive some attention. Although automatic text classification plays an important role in improving the browsability and accessibility of data, Electronic Theses and Dissertations (ETDs) have not received their fair share of attention, in spite of the huge number …


Identifying Facets Of Reader-Generated Online Reviews Of Children’S Books Based On A Textual Analysis Approach, Yunseon Choi, Soohyung Joo Jul 2020

Identifying Facets Of Reader-Generated Online Reviews Of Children’S Books Based On A Textual Analysis Approach, Yunseon Choi, Soohyung Joo

Information Science Faculty Publications

With the increasing popularity of social media, online reviews have become one of the primary information sources for book selection. Prior studies have analyzed online reviews, mostly in the domain of business. However, little research has examined the content of online book reviews of children’s books. Book reviews generated by book readers contain different aspects of information, such as opinions, feedback, or emotional responses, from the perspectives of readers. This study explores what aspects of the books are addressed in readers’ reviews, and then it intends to identify categorical features or facets of online book reviews of children’s books. We …


Scholarly Communication And Documentary Fragmentations In The Public Space: A Functional Citation Study, Fidelia Ibekwe, Lucie Loubère Dec 2019

Scholarly Communication And Documentary Fragmentations In The Public Space: A Functional Citation Study, Fidelia Ibekwe, Lucie Loubère

Proceedings from the Document Academy

This paper studies how academic content published in Open Edition.org, an online publication platform in the Social Sciences and Humanities is re-appropriated by members of the public. Our research is therefore concerned with the public appropriation of science and Open science. After extracting the contexts of citation of these content and mapping them, we propose a typology of citation functions as well as of citers (their origins and types). Our preliminary results indicated that academic literature is repurposed and cited by members of the public mainly as scientific warrant (support for their argumentation). We also found that academic content is …


Towards News Verification: Deception Detection Methods For News Discourse, Yimin Chen, Victoria L. Rubin, Niall Conroy Jan 2015

Towards News Verification: Deception Detection Methods For News Discourse, Yimin Chen, Victoria L. Rubin, Niall Conroy

FIMS Presentations

News verification is a process of determining whether a particular news report is truthful or deceptive. Deliberately deceptive (fabricated) news creates false conclusions in the readers’ minds. Truthful (authentic) news matches the writer’s knowledge. How do you tell the difference between the two in an automated way? To investigate this question, we analyzed rhetorical structures, discourse constituent parts and their coherence relations in deceptive and truthful news sample from NPR’s “Bluff the Listener”. Subsequently, we applied a vector space model to cluster the news by discourse feature similarity, achieving 63% accuracy. Our predictive model is not significantly better than chance …


Towards News Verification: Deception Detection Methods For News Discourse, Victoria Rubin, Niall Conroy, Yimin Chen Jan 2015

Towards News Verification: Deception Detection Methods For News Discourse, Victoria Rubin, Niall Conroy, Yimin Chen

Victoria Rubin

News verification is a process of determining whether a particular news report is truthful or deceptive. Deliberately deceptive (fabricated) news creates false conclusions in the readers’ minds. Truthful (authentic) news matches the writer’s knowledge. How do you tell the difference between the two in an automated way? To investigate this question, we analyzed rhetorical structures, discourse constituent parts and their coherence relations in deceptive and truthful news sample from NPR’s “Bluff the Listener”. Subsequently, we applied a vector space model to cluster the news by discourse feature similarity, achieving 63% accuracy. Our predictive model is not significantly better than chance …


Metadata And Linked Data In Word Sense Disambiguation, Matthew Corsmeier Jan 2015

Metadata And Linked Data In Word Sense Disambiguation, Matthew Corsmeier

Library Philosophy and Practice (e-journal)

Word Sense Disambiguation (WSD) can be assisted by taking advantage of the metadata embedded in the various ontologies, lexica, databases, etc… that exist in the Semantic Web. Automated processes that exploit the links already present in the Semantic Web can strengthen parsing of word senses by using user-contributed and semantically-linked data. These processes are only possible because of a commitment to interoperability and the creation of shared standards. This paper will review some of the most heavily used Linguistic Linked Open Data (LLOD) tools and models which show the most promise for using metadata to alleviate problems caused by polysemous …


Study Of Stemming Algorithms, Savitha Kodimala Dec 2010

Study Of Stemming Algorithms, Savitha Kodimala

UNLV Theses, Dissertations, Professional Papers, and Capstones

Automated stemming is the process of reducing words to their roots. The stemmed words are typically used to overcome the mismatch problems associated with text searching.


In this thesis, we report on the various methods developed for stemming. In particular, we show the effectiveness of n-gram stemming methods on a collection of documents.


Computational Linguistics For Metadata Building: Aggregating Text Processing Technologies For Enhanced Image Access, Judith Klavans, Carolyn Sheffield, Eileen Abels, Joan E. Beaudoin, Laura Jenemann, Jimmy Lin, Tom Lippincott, Rebecca Passonneau, Tandeep Sidhu, Dagobert Soergel, Tae Yano Aug 2008

Computational Linguistics For Metadata Building: Aggregating Text Processing Technologies For Enhanced Image Access, Judith Klavans, Carolyn Sheffield, Eileen Abels, Joan E. Beaudoin, Laura Jenemann, Jimmy Lin, Tom Lippincott, Rebecca Passonneau, Tandeep Sidhu, Dagobert Soergel, Tae Yano

School of Information Sciences Faculty Research Publications

We present a system which applies text mining using computational linguistic techniques to automatically extract, categorize, disambiguate and filter metadata for image access. Candidate subject terms are identified through standard approaches; novel semantic categorization using machine learning and disambiguation using both WordNet and a domain specific thesaurus are applied. The resulting metadata can be manually edited by image catalogers or filtered by semi-automatic rules. We describe the implementation of this workbench created for, and evaluated by, image catalogers. We discuss the system's current functionality, developed under the Computational Linguistics for Metadata Building (CLiMB) research project. The CLiMB Toolkit has been …