publications
* Denotes equal contribution
2023
- PyReason: Software for Open World Temporal LogicDyuman Aditya*, Kaustuv Mukerji*, Srikar Balasubramanian, Abhiraj Chaudhary, and Paulo ShakarianIn AAAI Spring Symposium, 2023
The growing popularity of neuro symbolic reasoning has led to the adoption of various forms of differentiable (i.e., fuzzy) first order logic. We introduce PyReason, a software framework based on generalized annotated logic that both captures the current cohort of differentiable logics and temporal extensions to support inference over finite periods of time with capabilities for open world reasoning. Further, PyReason is implemented to directly support reasoning over graphical structures (e.g., knowledge graphs, social networks, biological networks, etc.), produces fully explainable traces of inference, and includes various practical features such as type checking and a memory-efficient implementation. This paper reviews various extensions of generalized annotated logic integrated into our implementation, our modern, efficient Python-based implementation that conducts exact yet scalable deductive inference, and a suite of experiments. PyReason is available at: github.com/lab-v2/pyreason.
@inproceedings{aditya_pyreason_2023, author = {Aditya*, Dyuman and Mukerji*, Kaustuv and Balasubramanian, Srikar and Chaudhary, Abhiraj and Shakarian, Paulo}, title = {{PyReason}: Software for Open World Temporal Logic}, year = {2023}, booktitle = {{AAAI} Spring Symposium}, }
- Scalable Semantic Non-Markovian Simulation Proxy for Reinforcement LearningKaustuv Mukerji, Devendra Parkar, Lahiri Pokala, Dyuman Aditya, and Paulo ShakarianIn IEEE ICSC, 2023
Recent advances in reinforcement learning (RL) have shown much promise across a variety of applications. However, issues such as scalability, explainability, and Markovian assumptions limit its applicability in certain domains. We observe that many of these shortcomings emanate from the simulator as opposed to the RL training algorithms themselves. As such, we propose a semantic proxy for simulation based on a temporal extension to annotated logic. In comparison with two high-fidelity simulators, we show up to three orders of magnitude speed-up while preserving the quality of policy learned. In addition, we show the ability to model and leverage non-Markovian dynamics and instantaneous actions while providing an explainable trace describing the outcomes of the agent actions.
@inproceedings{mukherji2023scalable, author = {Mukerji, Kaustuv and Parkar, Devendra and Pokala, Lahiri and Aditya, Dyuman and Shakarian, Paulo}, title = {Scalable Semantic Non-Markovian Simulation Proxy for Reinforcement Learning}, year = {2023}, booktitle = {{IEEE} ICSC}, }