Synthesis via Input/Output Separation
08 Jul 2021
When: Thursday, July 8, 2021, 2:30pm (CEST).
Where: Zoom online, check the address in the Google Calendar Event.
Topic: Synthesis via Input/Output Separation.
Speaker: Dror Fried, Senior Lecturer at the Open
University of Israel.
Abstract
Decomposition is a commonly used technique in various algorithms in formal methods. When we observe systems that
describe a relation between inputs and outputs, a decomposition, or separation, of the system to input and output
components seems only natural. Specifically in synthesis, where the challenge is to construct a strategy or a function
that is essentially from inputs to outputs. But what sense does it really make? After all, the specification given by
the user, is intended to describe an involved relation between the input and output parts. As such, the algorithms
obtained so far are designed to exploit this tight correspondence between inputs and outputs towards strategy
generation. In this talk I show how we use input/output separation for synthesis in both areas of Boolean functional
synthesis and reactive synthesis. For Boolean functional synthesis I describe how to separate every Boolean formula to
input and output components. In reactive synthesis I describe applications in which it is natural for the designer to
write specifications that separate inputs and outputs. For both areas, I show how to exploit this separation towards
more efficient synthesis algorithms. This talk is based on joint works with Gal Amram, Suguman Bansal, Supratick
Chackraborty, Lucas Tabajara, Moshe Vardi, and Gera Weiss.
Short Bio
Dror Fried is a senior lecturer (assistant professor) in the Open University of Israel. He did his PhD in Ben-Gurion
University in artificial intelligence, and his postdoc in Rice University in formal methods with Moshe Vardi’s CAVR
group. His interests include formal methods, logic in computer science and automated reasoning.
Material
Synthesis From Temporal Specifications
24 Jun 2021
When: Thursday, June 24, 2021, 3pm (CEST).
Where: Zoom online, check the address in the Google Calendar Event.
Topic: Synthesis From Temporal Specifications.
Speaker: Nir Piterman, Associate Professor at the University of Gothenburg.
Abstract
In this talk I will present the GR(1) approach to synthesis, the automatic production of designs from their temporal
logic specifications. We are interested in reactive systems, systems that continuously interact with other programs,
users, or their environment and specifications in linear temporal logic. Classical solutions to synthesis use either two
player games or tree automata. I will give a short introduction to the technique of using two player games for
synthesis.
The classical solution to synthesis requires the usage of deterministic automata. This solution is 2EXPTIME-complete, is
quite complicated, and does not work well in practice. I will present a syntactic approach that restricts the kind of
properties users are allowed to write. It turns out that this approach is general enough and can be extended to cover
many properties written in practice.
I will survey some results in model-driven development and robot control.
Short Bio
Nir Piterman completed his PhD in 2005 at the Weizmann Institute of Science under the supervision of Amir Pnueli.
Between 2005-2007 he was a postdoc in Tom Henzinger’s group in the Ecoloe Polytechnique Federal de Lausanne. Between
2007-2010 he was a research Fellow in Imperial College London working with Michael Huth. Then, he joined the University
of Leicester in 2010 as a Lecturer, was promoted to Reader/Associate Professor in 2012. In March 2019 he joined the
Department of Computer Science and Engineering at the University of Gothenburg as a Universitets Lektor / Associate
Professor.
Material
- Video (Passcode: X?+96uYw)
Tractable Novelty Exploration over Continuous and Discrete Sequential Decision Problems
17 Jun 2021
When: Thursday, June 17, 2021, 12pm (CEST).
Where: Zoom online, check the address in the Google Calendar Event.
Topic: Tractable Novelty Exploration over Continuous and Discrete Sequential Decision Problems.
Speaker: Nir Lipovetzky, Senior Lecturer in Artificial Intelligence at the University
of Melbourne.
Abstract
Sequential decision problems, where an agent is trying to find a sequence of actions to maximise a utility function or
to satisfy a goal condition, have been the focus of different research communities. This talk focuses on the latest
advances over width-based planning algorithms, which have shown to yield state-of-the-art AI Planners that rely mostly
on structural exploration features, rather than goal-oriented heuristics.
Width-based planners search for a solution using a measure of the novelty of states, where states need to be defined
over a set of features. It is known that state novelty evaluation is exponential on the cardinality of the set of
features. In this talk, I will address two limitations of current width-based planning: 1) How to define state features
over continuous dynamics, where the space of features is unbounded, and 2) present new methods to obtain approximations
of novelty linear in the cardinality of the set of features instead of exponential.
I will discuss the performance of the resulting polynomial planners over discrete sequential decision problems (
classical planning) and compare over continuous problems with PPO, a state-of-the-art DRL algorithm. The continuous
problems are “classical control” benchmarks from openAI gym.
Short Bio
Nir Lipovetzky is a Senior Lecturer in Artificial Intelligence at the School of Computing and Information Systems, The
University of Melbourne. He’s a member of the Agent Lab group. He completed his PhD at the Artificial Intelligence and
Machine Learning Group, Universitat Pompeu Fabra, under the supervision of Prof. Hector Geffner. Then, he was a research
fellow for 3 years under the supervision of Prof. Peter Stuckey and Prof. Adrian Pearce, working on solving Mining
Scheduling problems through automated planning, constraint programming and operations research techniques.
His research focuses on how to introduce different approaches to the problem of inference in sequential decision
problems, as well as applications to autonomous systems.
Material
- Video (Passcode 7e5@pNu^)
Compositional Reinforcement Learning from Logical Specifications
10 Jun 2021
When: Thursday, June 10, 2021, 3pm (CEST).
Where: Zoom online, check the address in the Google Calendar Event.
Topic: Compositional Reinforcement Learning from Logical Specifications.
Speaker: Suguman Bansal, NSF/CRA Computing Innovation Postdoctoral Fellow at the University
of Pennsylvania.
Abstract
We study the problem of learning control policies for complex tasks given by logical specifications. Recent approaches
automatically generate a reward function from a given specification and use a suitable reinforcement learning algorithm
to learn a policy that maximizes the expected reward. These approaches, however, scale poorly to complex tasks that
require high-level planning. In this work, we develop a compositional learning approach, called DIRL, that interleaves
high-level planning and reinforcement learning. First, DIRL encodes the specification as an abstract graph; intuitively,
vertices and edges of the graph correspond to regions of the state space and simpler sub-tasks, respectively. Our
approach then incorporates reinforcement learning to learn neural network policies for each edge (sub-task) within a
Dijkstra-style planning algorithm to compute a high-level plan in the graph. An evaluation of the proposed approach on a
set of challenging control benchmarks with continuous state and action spaces demonstrates that it outperforms
state-of-the-art baselines.
This is joint work with Kishor Jothimurugan, Osbert Bastani, and Rajeev Alur.
Short Bio
Suguman Bansal is an NSF/CRA Computing Innovation Postdoctoral Fellow in the Department of Computer and Information
Sciences at the University of Pennsylvania. Her research interests lie at the intersection of Artificial Intelligence
and Formal Methods. Specifically, she works on developing tools and techniques for the automated synthesis of
computational systems that also offer formal guarantees of correctness.
She received her Ph.D. (2020) and M.S. (2016) in Computer Science from Rice University, and B.S. (with Honors) degree (
2014) in Mathematics and Computer Science from Chennai Mathematical Institute. She is the recipient of the NSF CI
Fellowship 2020, Future Faculty Fellowship 2019, EECS Rising Stars 2018, Andrew Ladd Fellowship 2016, and a Gold
Medal at the ACM Student Research Competition at POPL 2016.
Material
Natural Strategic Ability
03 Jun 2021
When: Thursday, June 3, 2021, 3pm (CEST).
Where: Zoom online, check the address in the Google Calendar Event.
Topic: Natural Strategic Ability.
Speaker: Vadim Malvone, Associate Professor in the Department of Computer
Science and Networks (INFRES) of Télécom Paris.
Abstract
In game theory, as well as in the semantics of game logics, a strategy can be represented by any function from states of
the game to the agent’s actions. That makes sense from the mathematical point of view, but not necessarily in the
context of human behavior. This is because humans are quite bad at executing complex plans, and also rather unlikely to
come up with such plans in the first place. In this work, we adopt the view of bounded rationality, and look only at “
simple” strategies in specifications of agents’ abilities. We formally define what “simple” means, and propose a variant
of alternating-time temporal logic that takes only such strategies into account. We also study the model checking
problem for the resulting semantics of ability.
Short Bio
Vadim Malvone is an Associate Professor in the Computer Science and Networks department (INFRES) of Télécom Paris. From November
2017 to June 2020, he was a postdoctoral researcher at the University of Evry under the supervision of Francesco
Belardinelli. In February 2018, he obtained my Ph.D. in Computer Science at the University of Naples “Federico II”, with
a thesis titled “Strategic Reasoning in Game Theory”, developed under the supervision of Aniello Murano. During his Ph.D.
program, he was a visiting researcher at the Polish Academy of Sciences, under the supervision of Wojtek Jamroga. In July
2014, Vadim obtained his Master’s degree in Computer Science, with a thesis titled “Graded modalities in strategic reasoning”
, developed under the supervision of Aniello Murano and Fabio Mogavero.
Material
A Unifying Framework for Observer-Aware Planning: Non-Markovian Reward Perspective
27 May 2021
When: Thursday, May 27, 2021, 3pm (CEST).
Where: Zoom online, check the address in the Google Calendar Event.
Topic: A Unifying Framework for Observer-Aware Planning: Non-Markovian Reward Perspective.
Speaker: Shuwa Miura, PhD student at the University of Massachusetts
Amherst, supervised by Shlomo Zilberstein.
Abstract
Being aware of observers and the inferences they make about an agent’s behavior is crucial for successful multi-agent
interaction. Existing works on observer-aware planning use different assumptions and techniques to produce
observer-aware behaviors. We introduce a unifying framework for producing observer-aware behaviors called Observer-Aware
MDP (OAMDP). We provide initial empirical evidence that OAMDPs can be used to improve interpretability of agent
behaviors and establish complexity results for polynomial-horizon OAMDPs. OAMDP defines a particular class of
non-Markovian Reward Decision Processes. Most previous work on non-Markovian reward have used a temporal logic over
finite histories to describe non-Markovian rewards. We show complexity results for polynomial-horizon non-Markovian
Reward Decision Processes with temporal logic, analogous to the results for OAMDP. We then discuss possible directions
for future research.