Flavors of UNCerTainty in Verification, Planning, and OpTimizatiON (FUNCTION)
This interdisciplinary workshop aims at bringing together researchers from different areas that formally model and analyse systems under uncertainty. We aim to solicit presentations and discussions on aspects of uncertainty including:
Such topics constitute active and relevant research in the areas including artificial intelligence, formal methods, operations research, and optimization. However, techniques and results have been obtained largely independently, and we expect strong synergy effects in sharing the specific yet often related views on uncertainty.
We ask for presentations of ongoing or previously published work to enable discussions on a broad range of topics. These presentations will not be subject to proceedings publication. We encourage all interested speakers to submit an abstract of maximum 1 page via Easychair:
EasyChair: Function2021 submission
Registration is free by contacting the workshop organizers or through:
Where workshop registration is free when registering for the ICALP conference and otherwise €10 until June 30 and €20 afterwards.
Eindhoven University of Technology, NL
University of York, UK
CISPA Helmholtz Center for Information Security in Saarbrücken, DE
University of California, Berkeley, US
University of Oxford, UK
University of Colorado Boulder, US
Eindhoven University of Technology, NL
Stanford University, US
Delft University of Technology, NL
Imperial College London, UK
11 July | |
---|---|
13:00 - 13:05 | Opening |
13:05 - 13:45 | Invited talk: Matthijs Spaan (Tu Delft) Safety in Online and Offline Reinforcement Learning |
13:45 - 14:00 | Regular talk: Tim Quatmann (RWTH Aachen University) Markov Models with Multiple Long-run Average and Total Reward Objectives |
14:00 - 14:15 | Regular talk: Marnix Suilen (Radboud University) Robust Policies for Uncertain POMDPs |
14:15 - 14:30 | Regular talk: Rania Taleb (Université du Québec à Chicoutimi), Raphael Khoury (Université du Québec à Chicoutimi) and Sylvain Hallé (Université du Québec à Chicoutimi) Runtime Verification Under Access Control Restrictions |
14:30 - 14:45 | Regular talk: Martin Fränzle (Carl von Ossietzky Universität Oldenburg) and Paul Kröger (Carl von Ossietzky Universität Oldenburg) Reconciling Formal Methods with Metrology |
14:45 - 15:00 | Sankalp Gilda (University of Florida) Uncertainty-Aware Learning for Improvements in Image Quality of the Canada-France-Hawaii Telescope |
15:00 - 15:30 | Break |
15:30 - 15:45 | Regular talk: Marcell Vazquez-Chanlatte (University of California, Berkeley), Sebastian Junges (University of California, Berkeley), Daniel J. Fremont (University of California, Santa Cruz) and Sanjit A. Seshia (University of California, Berkeley) Entropy-Guided Control Improvisation |
15:45 - 16:25 | Invited talk: Rayna Dimitrova (CISPA Helmholtz Center for Information Security in Saarbrücken) Probabilistic Hyperproperties of Markov Decision Processes |
16:25 - 16:40 | Regular talk: Hazem Torfah (University of California, Berkeley) Runtime Monitors for Markov Decision Processes |
16:40 - 17:20 | Invited talk: Morteza Lahijanian (University of Colorado Boulder) Data-driven Synthesis for Partially-known Stochastic Systems |
17:20 - 18:00 | Invited talk: Sebastian Junges Learning Memoryless Policies in POMDPs is ETR-complete. |
12 July | |
---|---|
13:00 - 13:40 | Invited talk: Rob Basten (TU Eindhoven) Achieving low downtime of high-tech systems with uncertain failure behaviour |
13:40 - 14:20 | Invited talk: Wolfram Wiesemann (Imperial College) Optimal Hospital Care Scheduling during the SARS-CoV2 Pandemic |
14:20 - 15:00 | Invited talk: Bruno Lacerda (Oxford Robotics Institute) Mission Planning for Robots in Partially Unknown Environments |
15:00 - 15:30 | Break |
15:30 - 16:10 | Invited talk: Ahmadreza Marandi (TU Eindhoven) Disjoint Bilinear Optimization: A Two-stage Robust Optimization Perspective |
16:10 - 16:30 | Regular talk: Nathanaël Fijalkow (CNRS, LaBRI, University of Bordeaux) Statistical comparison of algorithm performance through instance selection |
16:30 - 17:10 | Invited talk: Radu Calinescu (University of York) Probabilistic model checking: Where do the transition probabilities and rates of Markov models come from? |
17:10 - 17:50 | Invited talk: Ransalu Senanayake (Stanford Intelligent Systems Laboratory) Knowing the unknown unknowns in learning-based systems |
Matthijs Spaan
Safety in Online and Offline Reinforcement Learning
Abstract: Safety is a paramount issue for the application of reinforcement learning algorithms in the real world. We investigate safety from two perspectives, one where a reasonable performance must be ensured and another where safety constraints cannot be violated. In the first, the reinforcement learning agent only has access to a fix dataset of past trajectories and does not interact directly with the environment. Assuming the behavior policy that collected the data is available, the challenge is to compute a policy that outperforms the behavior policy. In this setting, we develop sample-efficient algorithms by exploiting the structure of the problem. In the second, the agent is trained by interacting directly with the environment, however there are safety constraints. Therefore, the agent cannot perform random exploration, which is typical in online reinforcement learning algorithms. Assuming the agent has access to an abstraction of the safety dynamics, we develop algorithms that can safely explore the environment and eventually converge to an optimal policy.
Wolfram Wiesemann
Optimal Hospital Care Scheduling During the SARS-CoV-2 Pandemic
Abstract: The COVID-19 pandemic has seen dramatic demand surges for hospital care that have placed a severe strain on health systems worldwide. As a result, policy makers are faced with the challenge of managing scarce hospital capacity so as to reduce the backlog of non-COVID patients whilst maintaining the ability to respond to any potential future increases in demand for COVID care. In this talk, we propose a nation-wide prioritization scheme that models each individual patient as a dynamic program whose states encode the patient’s health and treatment condition, whose actions describe the available treatment options, whose transition probabilities characterize the stochastic evolution of the patient’s health and whose rewards encode the contribution to the overall objectives of the health system. The individual patients’ dynamic programs are coupled through constraints on the available resources, such as hospital beds, doctors and nurses. We show that near-optimal solutions to the emerging weakly coupled counting dynamic program can be found through a fluid approximation that gives rise to a linear program whose size grows gracefully in the problem dimensions. Our case study for the National Health Service in England shows how years of life can be gained and costs reduced by prioritizing specific disease types over COVID patients, such as injury & poisoning, diseases of the respiratory system, diseases of the circulatory system, diseases of the digestive system and cancer.
Bruno Lacerda
Mission Planning for Robots in Partially Unknown Environments
Abstract: Robust mission planning algorithms for autonomous robots typically require explicit reasoning about the inherent uncertainty of environmental features. For example, service robots need to consider the behaviour of humans around them; autonomous underwater vehicles require models of underwater currents; and robots in a multi-robot system need to reason about how interaction with other robots can affect their performance. Oftentimes, these environmental features cannot be fully modelled before the system is deployed, and an open challenge in robotics is finding principled solutions that allow for robust deployments in such partially unknown environments. In this talk, I will describe our recent work on robust planning via regret minimisation of uncertain Markov decision processes (MDPs), an extension of MDPs where the transition and reward functions are not fully known. Then, assuming the robot is equipped with a sensor that can provide observations of the a priori unknown environmental features, I will describe our work on incorporating Gaussian processes with online MDP planning to enable robot deployments in partially unknown environments.
Ahmadreza Marandi
Disjoint Bilinear Optimization: A Two-stage Robust Optimization Perspective
Abstract: In this talk, we focus on a subclass of quadratic optimization problems, that is, disjoint bilinear optimization problems. We first show that disjoint bilinear optimization problems can be cast as two-stage robust linear optimization problems with fixed-recourse and right-hand-side uncertainty, which enables us to apply robust optimization techniques to solve the resulting problems. To this end, a solution scheme based on a blending of three popular robust optimization techniques is proposed. For disjoint bilinear optimization problems with a polyhedral feasible region and a general convex feasible region, we show that under mild regularity conditions, the convex relaxations of the original bilinear formulation and its two-stage robust reformulation obtained from a reformulation-linearization based technique and linear decision rules, respectively, are equivalent. For generic bilinear optimization problems, the convex relaxations from the reformulation-linearization based technique are generally tighter than the one from linear decision rules. Numerical experiments on bimatrix games and synthetic disjoint bilinear problem instances show the efficiency and effectiveness of the proposed solution scheme.
Radu Calinescu
Probabilistic model checking: Where do the transition probabilities and rates of Markov models come from?
Abstract: Probabilistic model checking is a powerful tool for establishing performance, dependability and other key properties of systems during design, verification, and at runtime. However, the usefulness of this tool depends on the accuracy of the Markov models being verified. In this talk, I will describe recent advances in exploiting data (available from offline testing, system logs, or runtime monitoring) to ensure that these models and their verification results are sufficiently accurate to support correct decision-making under epistemic uncertainty.
Rayna Dimitrova
Probabilistic Hyperproperties of Markov Decision Processes
Abstract: Trace properties, which are sets of execution traces, have been extensively used for specifying and reasoning about the correctness of systems. Trace properties, however, cannot describe information-flow security requirements, like noninterference, or important requirements relevant in the presence of uncertainty, like partial observation and robustness. Hyperproperties generalize trace properties and are capable of describing requirements that relate multiple system executions. Over the past decade, a rich set of temporal logics for the specification of hyperproperties and techniques for reasoning about them have been developed. In this talk I will present a recently introduced temporal logic for the specification of hyperproperties of Markov decision processes (MDPs), called Probabilistic Hyper Logic (PHL). PHL extends classic probabilistic logics with quantification over schedulers and traces. It can express a wide range of hyperproperties for probabilistic systems, including both classical applications, such as probabilistic noninterference and differential privacy, as well as novel applications in areas such as planning. A consequence of the generality of the logic is that the model checking problem for PHL is undecidable. I will present methods both for proving and for refuting formulas from a fragment of the logic that includes many probabilistic hyperproperties of interest. I will discuss how probabilistic hyperproperties can be used to specify and reason about the effects of uncertainty. This talk is based on joint work with Bernd Finkbeiner and Hazem Torfah.
Rob Basten
Achieving low downtime of high-tech systems with uncertain failure behaviour
Abstract: High-tech systems, such as wafer steppers, baggage handling systems or industrial printers, are expensive and complex. This means that their users do not typically invest in redundancy and thus require low downtime and high reliability. At the same time, it is very hard to predict when something in the system may fail, and what will fail. This means that it is hard to determine when to perform preventive maintenance to prevent failures from occurring, and which spare parts to stock to quickly perform corrective maintenance after a failure has occurred. We will discuss the challenges in maintenance optimization and spare parts inventory control and we will see how the research in operations management and industrial engineering is moving from assuming known failure behaviour to acknowledging that it may not even be clear what type of distribution fits the failure behaviour of the components in the system. It will become clear that if we want to increase our impact in practice, much (joint) research is required from researchers in operations management, industrial engineering, and computer science.
Tim Quatmann
Markov Models with Multiple Long-run Average and Total Reward Objectives
Abstract: We present an efficient procedure for multi-objective model checking of long-run average (LRA) reward and total reward objectives as well as their combination. We consider this for Markov automata (MAs), a compositional model that captures both traditional Markov decision processes (MDPs) as well as a continuous-time variant thereof. Our approach thus allows us to study tradeoffs between potentially conflicting objectives of systems involving controlled- and randomized decisions as well as random timings given by exponential distributions.
Marcell Vazquez-Chanlatte, Sebastian Junges, Daniel J. Fremont and Sanjit A. Seshia
Entropy-Guided Control Improvisation
Abstract: High level declarative constraints provide a powerful (and popular) way to define and construct control policies; however, most synthesis algorithms do not support specifying the degree of randomness (unpredictability) of the resulting controller. In many contexts, e.g., patrolling, testing, behavior prediction,and planning on idealized models, predictable or biased controllers are undesirable. To address these concerns, we introduce the Entropic Reactive Control Improvisation (ERCI) framework and algorithm which supports synthesizing control policies for stochastic games that are declaratively specified by (i) a \emph{hard constraint} specifying what must occur, (ii) a \emph{soft constraint} specifying what typically occurs, and (iii) a randomization constraint specifying the unpredictability and variety of the controller, as quantified using causal entropy. This framework, extends the state of the art by supporting arbitrary combinations of adversarial and probabilistic uncertainty in the environment. ERCI enables a flexible modeling formalism which we argue, theoretically and empirically, remains tractable.
Marnix Suilen
Robust Policies for Uncertain POMDPs
Abstract: We study the problem of computing robust finite-memory policies for uncertain partially observable Markov decision processes (POMDPs). Uncertain POMDPs (uPOMDPs) form an extension of the standard POMDP model where the transition or observation probabilities are not exactly known, but given by an uncertainty set, like intervals of probabilities or likelihood functions. Such uncertainty sets can, for instance, account for sensor imprecisions or human behavior. We include such uncertainties in the decision-making by computing finite-memory policies that satisfy a reachability or expected reward specification under the worst-case instance of the uncertainty in the model. Experiments show that our method based on convex optimization is feasible for uPOMDPs of several hundreds of thousands of states. We show the applicability of our approach on the example case study of spacecraft motion planning.
Rania Taleb, Raphael Khoury and Sylvain Hallé
Runtime Verification Under Access Control Restrictions
Abstract: We define a logical framework that permits runtime verification to take place when a monitor has incomplete or uncertain information about the underlying trace. Uncertainty is modeled as a stateful access control proxy that has the capacity to turn events into sets of possible events, resulting in what we call a “multi-trace”. We describe a model of both proxy and monitor as extensions of Mealy machines, and provide an algorithm to lift a classical monitor into a sound, loss-tolerant monitor. Experiments on various scenarios show that the approach can account for various types of data degradation and access limitations, provides a tighter verdict than existing works in some cases.
Morteza Lahijanian
Data-driven Synthesis for Partially-known Stochastic Systems
Abstract: Cyber-physical systems (CPS) are widely used in our society, from smart grids to autonomous cars. They are rapidly gaining more capabilities via AI components and serving in ever more safety-critical roles. However, as their complexities grow, the more uncertain their models become, and hence, the more difficult it becomes to provide safety guarantees for their performance. In this talk, I argue that formal methods combined with machine learning techniques can form powerful tools to address this problem. To this end, I present a data-driven framework for correct-by-construction control synthesis for partially-known stochastic CPS. The framework takes a set of input-output data and a temporal logic specification as input and returns a control strategy under which the system is guaranteed to satisfy the specification. Underlying this approach is a finite abstraction paradigm that leverages Gaussian process regression and uncertain Markov processes. The abstraction accounts for both the stochastic behavior of the system and the uncertainty in the regression step. Then, using automaton-based reasoning and optimization techniques, a control strategy is synthesized that maximizes the satisfaction probability of the specification and is robust against all the uncertainties in the abstraction. I show that this strategy is near-optimal and provide a bound on its distance (error) to the optimal controller. I also present a series of case studies, including both (unknown) linear and non-linear switched stochastic systems, which validate the theoretical results.
Ransalu Senanayake
Knowing the unknown unknowns in learning-based systems
Abstract: Autonomous agents have already gained the capability to perform individual tasks such as object detection and lane following, especially in simple, static environments. However, for robots to safely operate in the real-world, it is vital for them to quantify the uncertainty around them. In this talk, I will talk about how we can leverage tools from approximate Bayesian inference to estimate multimodal aleatoric and epistemic uncertainty in various learning-based systems with applications to robot perception and autonomous driving.
Bio: Ransalu Senanayake is a postdoctoral research scholar in Artificial Intelligence at the Stanford Intelligent Systems Laboratory, Stanford University. He is also a member of the Stanford Center for AI Safety and the SAIL-Toyota Center for AI Research. Prior to joining Stanford, Ransalu obtained a PhD in Computer Science from the University of Sydney in 2019, an MPhil in Industrial Engineering and Decision Analytics from the Hong Kong University of Science and Technology in 2013, and a BEng Honours in Electronic Engineering from the Sheffield Hallam University in 2011. He is an Associate Editor for the IEEE International Conference on Intelligent Robots and Systems (IROS) in 2021.