Search Results for author: Makai Mann

Found 5 papers, 1 papers with code

Safety-Aware Task Composition for Discrete and Continuous Reinforcement Learning

no code implementations29 Jun 2023 Kevin Leahy, Makai Mann, Zachary Serlin

We advance the state of the art in Boolean composition of learned tasks with three contributions: i) introduce two distinct notions of safety in this framework; ii) show how to enforce either safety semantics, prove correctness (under some assumptions), and analyze the trade-offs between the two safety notions; and iii) extend Boolean composition from discrete action spaces to continuous action spaces.

reinforcement-learning Reinforcement Learning (RL) +1

STL: Surprisingly Tricky Logic (for System Validation)

no code implementations26 May 2023 Ho Chit Siu, Kevin Leahy, Makai Mann

The ground-truth validity of a specification, subjects' familiarity with formal methods, and subjects' level of education were found to be significant factors in determining validation correctness.

valid

Lightweight Online Learning for Sets of Related Problems in Automated Reasoning

1 code implementation18 May 2023 Haoze Wu, Christopher Hahn, Florian Lonsing, Makai Mann, Raghuram Ramanujan, Clark Barrett

We present Self-Driven Strategy Learning ($\textit{sdsl}$), a lightweight online learning methodology for automated reasoning tasks that involve solving a set of related problems.

Uncertainty Quantification for Recursive Estimation in Adaptive Safety-Critical Control

no code implementations4 Apr 2023 Max H. Cohen, Makai Mann, Kevin Leahy, Calin Belta

In this paper, we present a framework for online parameter estimation and uncertainty quantification in the context of adaptive safety-critical control.

Uncertainty Quantification

Learning Minimally-Violating Continuous Control for Infeasible Linear Temporal Logic Specifications

no code implementations3 Oct 2022 Mingyu Cai, Makai Mann, Zachary Serlin, Kevin Leahy, Cristian-Ioan Vasile

This is achieved by decomposing an infeasible LTL formula into several reach-avoid sub-tasks with shorter horizons, which can be trained in a modular DRL architecture.

Continuous Control

Cannot find the paper you are looking for? You can Submit a new open access paper.