PublicationsYevgeniy Vorobeychik logo

2024

Junlin Wu, Jiongxiao Wang, Chaowei Xiao, Chenguang Wang, Ning Zhang, and Yevgeniy Vorobeychik. Preference Poisoning Attacks on Reward Model Learning. (SP 2025), to appear.

Junlin Wu, Huan Zhang, Yevgeniy Vorobeychik. Verified Safe Reinforcement Learning for Neural Network Dynamic Models. (NeurIPS 2024), to appear.

Luise Ge, Daniel Halpern, Evi Micha, Ariel D. Procaccia, Itai Shapira, Yevgeniy Vorobeychik, Junlin Wu. Axioms for AI Alignment from Human Feedback. (NeurIPS 2024), to appear.

Anindya Sarkar, Srikumar Sastry, Aleksis Pirinen, Chongjie Zhang, Nathan Jacobs, Yevgeniy Vorobeychik. GOMAA-Geo: GOal Modality Agnostic Active Geo-localization. (NeurIPS 2024), to appear.

Jiongxiao Wang, Junlin Wu, Muhao Chen, Yevgeniy Vorobeychik, and Chaowei Xiao. On the Exploitability of Reinforcement Learning with Human Feedback for Large Language Models. [code] (ACL 2024).

Taha Eghtesad, Sirui Li, Yevgeniy Vorobeychik, and Aron Laszka. Multi-Agent Reinforcement Learning for Assessing False-Data Injection Attacks on Transportation Networks. [code] (AAMAS 2024).

Andrew Estornell, Tina Zhang, Sanmay Das, Chien-Ju Ho, Brendan Juba, and Yevgeniy Vorobeychik. The Impact of Features Used by Algorithms on Perceptions of Fairness. (IJCAI 2024).

Jayanth Yetukuri, Ian Hardy, Yevgeniy Vorobeychik, Berk Ustun, Yang Liu. Providing Fair Recourse over Plausible Groups. (AAAI 2024).

Anindya Sarkar, Michael Lanier, Scott Alfeld, Jiarui Feng, Roman Garnett, Nathan Jacobs, and Yevgeniy Vorobeychik. A Visual Active Search Framework for Geospatial Exploration. [code] (WACV 2024).

Michael Lanier, Ying Xu, Nathan Jacobs, Chongjie Zhang, and Yevgeniy Vorobeychik. Learning Interpretable Policies in Hindsight-Observable POMDPs through Partially Supervised Reinforcement Learning. (ICMLA 2024), to appear.

James Thomas Brown, Ellen Wright Clayton, Michael E. Matheny, Murat Kantarcioglu, Yevgeniy Vorobeychik, Bradley A. Malin: Robin Hood: A De-identification Method to Preserve Minority Representation for Disparities Research. (PSD 2024: 67-83).

Junlin Wu, Hussein Sibai, Yevgeniy Vorobeychik: Certifying Safety in Reinforcement Learning under Adversarial Perturbation Attacks. (DLSP 2024).

Xia Li, Andrea L. Bertozzi, P. Jeffrey Brantingham, and Yevgeniy Vorobeychik. Optimal policy for control of epidemics with constrained time intervals and region-based interactions. (AIMS Networks and Heterogeneous Media 2024).

Benjamin Miller, Zohair Shafi, Wheeler Ruml, Yevgeniy Vorobeychik, Tina Eliassi-Rad and Scott Alfeld. Attacking shortest paths by cutting edges. [code] (TKDD 2024).

© 2024 Yevgeniy Vorobeychik