Open Access. Powered by Scholars. Published by Universities.®
- Publication Type
Articles 1 - 9 of 9
Full-Text Articles in Computer Engineering
Decision Making For Dynamic Systems Under Uncertainty: Predictions And Parameter Recomputations, Leobardo Valera
Decision Making For Dynamic Systems Under Uncertainty: Predictions And Parameter Recomputations, Leobardo Valera
Open Access Theses & Dissertations
In this Thesis, we are interested in making decision over a model of a dynamic system. We want to know, on one hand, how the corresponding dynamic phenomenon unfolds under different input parameters (simulations). These simulations might help researchers to design devices with a better performance than the actual ones. On the other hand, we are also interested in predicting the behavior of the dynamic system based on knowledge of the phenomenon in order to prevent undesired outcomes. Finally, this Thesis is concerned with the identification of parameters of dynamic systems that ensure a specific performance or behavior.
Understanding the …
Optimizing Computer Representation And Computer Processing Of Epistemic Uncertainty For Risk-Informed Decision Making: Finances Etc., Vladik Kreinovich, Nitaya Buntao, Olga Kosheleva
Optimizing Computer Representation And Computer Processing Of Epistemic Uncertainty For Risk-Informed Decision Making: Finances Etc., Vladik Kreinovich, Nitaya Buntao, Olga Kosheleva
Departmental Technical Reports (CS)
Uncertainty is usually gauged by using standard statistical characteristics: mean, variance, correlation, etc. Then, we use the known values of these characteristics (or the known bounds on these values) to select a decision. Sometimes, it becomes clear that the selected characteristics do not always describe a situation well; then other known (or new) characteristics are proposed. A good example is description of volatility in finance: it started with variance, and now many descriptions are competing, all with their own advantages and limitations.
In such situations, a natural idea is to come up with characteristics tailored to specific application areas: e.g., …
Estimating Information Amount Under Uncertainty: Algorithmic Solvability And Computational Complexity, Vladik Kreinovich, Gang Xiang
Estimating Information Amount Under Uncertainty: Algorithmic Solvability And Computational Complexity, Vladik Kreinovich, Gang Xiang
Departmental Technical Reports (CS)
Measurement results (and, more generally, estimates) are never absolutely accurate: there is always an uncertainty, the actual value x is, in general, different from the estimate X. Sometimes, we know the probability of different values of the estimation error dx=X-x, sometimes, we only know the interval of possible values of dx, sometimes, we have interval bounds on the cdf of dx. To compare different measuring instruments, it is desirable to know which of them brings more information - i.e., it is desirable to gauge the amount of information. For probabilistic uncertainty, this amount of information is described by Shannon's entropy; …
Optimal Sensor Placement In Environmental Research: Designing A Sensor Network Under Uncertainty, Aline James, Craig Tweedie, Tanja Magoc, Vladik Kreinovich, Martine Ceberio
Optimal Sensor Placement In Environmental Research: Designing A Sensor Network Under Uncertainty, Aline James, Craig Tweedie, Tanja Magoc, Vladik Kreinovich, Martine Ceberio
Departmental Technical Reports (CS)
One of our main challenges in meteorology and environment research is that in many important remote areas, sensor coverage is sparse, leaving us with numerous blind spots. Placement and maintenance of sensors in these areas are expensive. It is therefore desirable to find out how, within a given budget, we can design a sensor network are important activities was developing reasonable techniques for sensor that would provide us with the largest amount of useful information while minimizing the size of the "blind spot" areas which is not covered by the sensors.
This problem is very difficult even to formulate in …
Model Fusion Under Probabilistic And Interval Uncertainty, With Application To Earth Sciences, Omar Ochoa, Aaron A. Velasco, Christian Servin, Vladik Kreinovich
Model Fusion Under Probabilistic And Interval Uncertainty, With Application To Earth Sciences, Omar Ochoa, Aaron A. Velasco, Christian Servin, Vladik Kreinovich
Departmental Technical Reports (CS)
One of the most important studies of the earth sciences is that of the Earth's interior structure. There are many sources of data for Earth tomography models: first-arrival passive seismic data (from the actual earthquakes), first-arrival active seismic data (from the seismic experiments), gravity data, and surface waves. Currently, each of these datasets is processed separately, resulting in several different Earth models that have specific coverage areas, different spatial resolutions and varying degrees of accuracy. These models often provide complimentary geophysical information on earth structure (P and S wave velocity structure).
Combining the information derived from each requires a joint …
Maximum Entropy In Support Of Semantically Annotated Datasets, Paulo Pinheiro Da Silva, Vladik Kreinovich, Christian Servin
Maximum Entropy In Support Of Semantically Annotated Datasets, Paulo Pinheiro Da Silva, Vladik Kreinovich, Christian Servin
Departmental Technical Reports (CS)
One of the important problems of semantic web is checking whether two datasets describe the same quantity. The existing solution to this problem is to use these datasets' ontologies to deduce that these datasets indeed represent the same quantity. However, even when ontologies seem to confirm the identify of the two corresponding quantities, it is still possible that in reality, we deal with somewhat different quantities. A natural way to check the identity is to compare the numerical values of the measurement results: if they are close (within measurement errors), then most probably we deal with the same quantity, else …
Propagation And Provenance Of Probabilistic And Interval Uncertainty In Cyberinfrastructure-Related Data Processing And Data Fusion, Paulo Pinheiro Da Silva, Aaron A. Velasco, Martine Ceberio, Christian Servin, Matthew G. Averill, Nicholas Ricky Del Rio, Luc Longpre, Vladik Kreinovich
Propagation And Provenance Of Probabilistic And Interval Uncertainty In Cyberinfrastructure-Related Data Processing And Data Fusion, Paulo Pinheiro Da Silva, Aaron A. Velasco, Martine Ceberio, Christian Servin, Matthew G. Averill, Nicholas Ricky Del Rio, Luc Longpre, Vladik Kreinovich
Departmental Technical Reports (CS)
In the past, communications were much slower than computations. As a result, researchers and practitioners collected different data into huge databases located at a single location such as NASA and US Geological Survey. At present, communications are so much faster that it is possible to keep different databases at different locations, and automatically select, transform, and collect relevant data when necessary. The corresponding cyberinfrastructure is actively used in many applications. It drastically enhances scientists' ability to discover, reuse and combine a large number of resources, e.g., data and services.
Because of this importance, it is desirable to be able to …
A New Cauchy-Based Black-Box Technique For Uncertainty In Risk Analysis, Vladik Kreinovich, Scott Ferson
A New Cauchy-Based Black-Box Technique For Uncertainty In Risk Analysis, Vladik Kreinovich, Scott Ferson
Departmental Technical Reports (CS)
Uncertainty is very important in risk analysis. A natural way to describe this uncertainty is to describe a set of possible values of each unknown quantity (this set is usually an interval), plus any additional information that we may have about the probability of different values within this set. Traditional statistical techniques deal with the situations in which we have a complete information about the probabilities; in real life, however, we often have only partial information about them. We therefore need to describe methods of handling such partial information in risk analysis. Several such techniques have been presented, often on …
Why 95% And Two Sigma? A Theoretical Justification For An Empirical Measurement Practice, Hung T. Nguyen, Vladik Kreinovich, Chin-Wang Tao
Why 95% And Two Sigma? A Theoretical Justification For An Empirical Measurement Practice, Hung T. Nguyen, Vladik Kreinovich, Chin-Wang Tao
Departmental Technical Reports (CS)
The probability p(k) that the value of a random variable is far away from the mean (e.g. further than k standard deviations away) is so small that this possibility can be often safely ignored. It is desirable to select k for which the dependence of the probability p(k) on the distribution is the smallest possible. Empirically, this dependence is the smallest for k between 1.5 and 2.5. In this paper, we give a theoretical explanation for this empirical result.