Open Access. Powered by Scholars. Published by Universities.®

Engineering Commons

Open Access. Powered by Scholars. Published by Universities.®

Series

Computer Engineering

University of Louisville

AI safety

Publication Year

Articles 1 - 5 of 5

Full-Text Articles in Engineering

Understanding And Avoiding Ai Failures: A Practical Guide, Robert Williams, Roman Yampolskiy Sep 2021

Understanding And Avoiding Ai Failures: A Practical Guide, Robert Williams, Roman Yampolskiy

Faculty Scholarship

As AI technologies increase in capability and ubiquity, AI accidents are becoming more common. Based on normal accident theory, high reliability theory, and open systems theory, we create a framework for understanding the risks associated with AI applications. This framework is designed to direct attention to pertinent system properties without requiring unwieldy amounts of accuracy. In addition, we also use AI safety principles to quantify the unique risks of increased intelligence and human-like qualities in AI. Together, these two fields give a more complete picture of the risks of contemporary AI. By focusing on system properties near accidents instead of …


Transdisciplinary Ai Observatory—Retrospective Analyses And Future-Oriented Contradistinctions, Nadisha Marie Aliman, Leon Kester, Roman Yampolskiy Jan 2021

Transdisciplinary Ai Observatory—Retrospective Analyses And Future-Oriented Contradistinctions, Nadisha Marie Aliman, Leon Kester, Roman Yampolskiy

Faculty Scholarship

In the last years, artificial intelligence (AI) safety gained international recognition in the light of heterogeneous safety-critical and ethical issues that risk overshadowing the broad beneficial impacts of AI. In this context, the implementation of AI observatory endeavors represents one key research direction. This paper motivates the need for an inherently transdisciplinary AI observatory approach integrating diverse retrospective and counterfactual views. We delineate aims and limitations while providing hands-on-advice utilizing concrete practical examples. Distinguishing between unintentionally and intentionally triggered AI risks with diverse socio-psycho-technological impacts, we exemplify a retrospective descriptive analysis followed by a retrospective counterfactual risk analysis. Building on …


Chess As A Testing Grounds For The Oracle Approach To Ai Safety, James D. Miller, Roman Yampolskiy, Olle Häggström, Stuart Armstrong Sep 2020

Chess As A Testing Grounds For The Oracle Approach To Ai Safety, James D. Miller, Roman Yampolskiy, Olle Häggström, Stuart Armstrong

Faculty Scholarship

To reduce the danger of powerful super-intelligent AIs, we might make the first such AIs oracles that can only send and receive messages. This paper proposes a possibly practical means of using machine learning to create two classes of narrow AI oracles that would provide chess advice: those aligned with the player's interest, and those that want the player to lose and give deceptively bad advice. The player would be uncertain which type of oracle it was interacting with. As the oracles would be vastly more intelligent than the player in the domain of chess, experience with these oracles might …


Towards Ai Welfare Science And Policies, Soenke Ziesche, Roman Yampolskiy Mar 2019

Towards Ai Welfare Science And Policies, Soenke Ziesche, Roman Yampolskiy

Faculty Scholarship

In light of fast progress in the field of AI there is an urgent demand for AI policies. Bostrom et al. provide “a set of policy desiderata”, out of which this article attempts to contribute to the “interests of digital minds”. The focus is on two interests of potentially sentient digital minds: to avoid suffering and to have the freedom of choice about their deletion. Various challenges are considered, including the vast range of potential features of digital minds, the difficulties in assessing the interests and wellbeing of sentient digital minds, and the skepticism that such research may encounter. Prolegomena …


A Psychopathological Approach To Safety Engineering In Ai And Agi, Vahid Behzadan, Arslan Munir, Roman V. Yampolskiy Aug 2018

A Psychopathological Approach To Safety Engineering In Ai And Agi, Vahid Behzadan, Arslan Munir, Roman V. Yampolskiy

Faculty Scholarship

The complexity of dynamics in AI techniques is already approaching that of complex adaptive systems, thus curtailing the feasibility of formal controllability and reachability analysis in the context of AI safety. It follows that the envisioned instances of Artificial General Intelligence (AGI) will also suffer from challenges of complexity. To tackle such issues, we propose the modeling of deleterious behaviors in AI and AGI as psychological disorders, thereby enabling the employment of psychopathological approaches to analysis and control of misbehaviors. Accordingly, we present a discussion on the feasibility of the psychopathological approaches to AI safety, and propose general directions for …