| | |
461 | Leveraging Score-based Models for Generating Penalization in Model-based Offline Reinforcement Learning | Zeyuan Liu, Zhirui Fang, Jiafei Lyu, Xiu Li |
517 | Minimizing Rosenthal’s Potential in Monotone Congestion Games | Vittorio Bilò, Angelo Fanelli, Laurent Gourvès, Christos Tsoufis, Cosimo Vinci |
284 | Approximation Algorithms for Connected Maximum Coverage | Gianlorenzo D’Angelo, Esmaeil Delfaraz |
973 | Hierarchical Imitation Learning of Team Behavior from Heterogeneous Demonstrations | Sangwon Seo, Vaibhav V. Unhelkar |
120 | Safe Pareto Improvements for Expected Utility Maximizers in Program Games | Anthony Digiovanni, Jesse Clifton, Nicolas Macé |
187 | ACORN: Acyclic Coordination with Reachability Network to Reduce Communication Redundancy in Multi-Agent Systems | Xie Yi, Ziqing Zhou, Chun Ouyang, Siao Liu, Linqiang Hu, Zhongxue Gan |
213 | Hypothesis-Driven Explainable Goal Recognition | Abeer Alshehri, Hissah Alotaibi, Tim Miller, Mor Vered |
932 | Coherence-Driven Multimodal Safety Dialogue with Active Learning for Embodied Agents | Sabit Hassan, Hye-Young Chung, Xiang Zhi Tan, Malihe Alikhani |
727 | Asymptotic Existence of Class Envy-free Matchings | Tomohiko Yokoyama, Ayumi Igarashi |
790 | Learning Real-Life Approval Elections | Piotr Faliszewski, Łukasz Janeczko, Andrzej Kaczmarczyk, Marcin Kurdziel, Grzegorz Pierczyński, Stanisław Szufa |
174 | Probably Correct Optimal Stable Matching for Two-Sided Market Under Uncertainty | Andreas Athanasopoulos, Anne-Marie George, Christos Dimitrakakis |
1107 | Reputation-Filtered Reward Reshaping: Encouraging Cooperation in High Dimensional Semi-Cooperative Multi-agent Settings | Hassan Raissouni, Wissal Bekhti, Btissam El Khamlichi, Amal Seghrouchni |
98 | MacLight: Multi-scene Aggregation Convolutional Learning for Traffic Signal Control | Sunbowen Lee, Hongqin Lyu, Yicheng Gong, Sun Yingying, Chao Deng |
367 | A Scoresheet for Explainable AI | Michael Winikoff, John Thangarajah, Sebastian Rodriguez |
1176 | Hitchhiker’s Guide to Patrolling: Path-Finding for Energy-Sharing Drone-UGV Teams | Jonathan Diller, Qi Han, Robert Byers, James Dotterweich, James Humann |
47 | Azorus: Commitments over Protocols for BDI Agents | Amit K. Chopra, Matteo Baldoni, Samuel H. Christie V, Munindar P. Singh |
314 | Soft Condorcet Optimization for Ranking of General Agents | Marc Lanctot, Kate Larson, Michael Kaisers, Quentin Berthet, Ian Gemp, Manfred Diaz, Roberto-Rafael Maura-Rivero, Yoram Bachrach, Anna Koop, Doina Precup |
363 | On the Complexity of Learning to Cooperate in Populations of Socially Rational Agents | Saptarashmi Bandyopadhyay, Mustafa Mert Çelikok, Robert Loftin |
1209 | The Degree of (Extended) Justified Representation and Its Optimization | Biaoshuai Tao, Chengkai Zhang, Houyu Zhou |
1010 | Finite-Horizon Single-Pull Restless Bandits: An Efficient Index Policy For Scarce Resource Allocation | Guojun Xiong, Haichuan Wang, Yuqi Pan, Saptarshi Mandal, Sanket Shah, Niclas Boehmer, Milind Tambe |
699 | Explaining Facial Expression Recognition | Sanjeev Nahulanthran, Leimin Tian, Dana Kulic, Mor Vered |
1351 | The effect of agent-based feedback on prosociality in social dilemmas | Jennifer Renoux, Filipa Correia, Joana Campos, Lucas Morillo-Mendez, Neziha Akalin, Fernando P. Santos, Ana Paiva |
614 | Approximation Ratio for Preference Aggregation Using Tree CP-Nets | Abu Mohammad Hammad Ali, Daniel Ogundare, Boting Yang, Sandra Zilles |
28 | Game-Theoretically Secure Distributed Protocols for Fair Allocation in Coalitional Games | T-H. Hubert Chan, Qipeng Kuang, Quan Xue |
992 | Large Language Models for Virtual Human Gesture Selection | Parisa Ghanad Torshizi, Laura B. Hensel, Ari Shapiro, Stacy Marsella |
239 | Self-Supervised Multi-Agent Diversity with Nonparametric Entropy Maximization | Tianxu Li, Kun Zhu |
1142 | Who Reviews The Reviewers? A Multi-Level Jury Problem | Ben Abramowitz, Omer Lev, Nicholas Mattei |
302 | Divide and Conquer: Provably Unveiling the Pareto Front with Multi-Objective Reinforcement Learning | Willem Röpke, Mathieu Reymond, Patrick Mannion, Diederik M Roijers, Ann Nowé, Roxana Rădulescu |
683 | Data Pricing for Graph Neural Networks without Pre-purchased Inspection | Yiping Liu, Mengxiao Zhang, Jiamou Liu, Song Yang |
149 | Goal Recognition via Variational Causality | Jiaqi Wen, Leonardo Rosa Amado |
521 | Timed Obstruction Logic: A Timed Approach to Dynamic Game Reasoning | James Ortiz, Vadim Malvone, Jean Leneutre |
499 | The Many Challenges of Human-Like Agents in Virtual Game Environments | Maciej Świechowski, Dominik Slezak |
622 | Scalable Offline Reinforcement Learning for Mean Field Games | Axel Brunnbauer, Julian Lemmel, Zahra Babaiee, Sophie A. Neubauer, Radu Grosu |
212 | Teamwork Makes the Defense Work: Comprehensive Vulnerability Defense Resource Allocation | Siyu Liu, Rida Bazzi, Fei Fang, Tiffany Bao |
45 | Game Theory with Simulation in the Presence of Unpredictable Randomisation | Vojtech Kovarik, Nathaniel Sauerberg, Lewis Hammond, Vincent Conitzer |
612 | In-context Learning from Language Models can Improve Embodied Instruction-following | Pengyuan Wang, Jing-Cheng Pang, Wang Chenyang, Xu-Hui Liu, Tian-Shuo Liu, Si-Hang Yang, Yang Yu, Hong Qian |
771 | A View of the Certainty-Equivalence Method for PAC RL as an Application of the Trajectory Tree Method | Shivaram Kalyanakrishnan, Sheel Shah, Santhosh Kumar Guguloth |
550 | Ensemble Value Functions for Efficient Exploration in Multi-Agent Reinforcement Learning | Lukas Schäfer, Oliver Slumbers, Stephen Marcus Mcaleer, Yali Du, Stefano V Albrecht, David Henry Mguni |
1269 | Geometric Freeze-Tag Problem | Sharareh Alipour, Kajal Baghestani, Mahdis Mirzaei, Soroush Sahraei |
966 | Reinforcement Learning-based Approach for Vehicle-to-Building Charging with Heterogeneous Agents and Long Term Rewards | Fangqi Liu, Rishav Sen, Jose Paolo Talusan, Ava Pettet, Aaron Kandel, Yoshinori Suzue, Ayan Mukhopadhyay, Abhishek Dubey |
108 | Gricean Norms as a Basis for Effective Collaboration | Fardin Saad, Pradeep K. Murukannaiah, Munindar P. Singh |
43 | ReSCOM: Reward-Shaped Curriculum for Efficient Multi-Agent Communication Learning | Xinghai Wei, Tingting Yuan, Jie Yuan, Dongxiao Liu, Xiaoming Fu |
1381 | Modeling the Centaur: Human-Machine Synergy in Sequential Decision Making | David Shoresh, Yonatan Loewenstein |
605 | An Organizationally-Oriented Approach to Enhancing Explainability and Control in Multi-Agent Reinforcement Learning | Julien Soulé, Jean-Paul Jamont, Michel Occello, Louis-Marie Traonouez, Paul Théron |
31 | Approximating One-Sided and Two-Sided Nash Social Welfare With Capacities | Salil Gokhale, Harshul Sagar, Rohit Vaish, Jatin Yadav |
1109 | Game of Thoughts: Iterative Reasoning in Game-Theoretic Domains with Large Language Models | Benjamin Kempinski, Ian Gemp, Kate Larson, Yoram Bachrach, Marc Lanctot, Tal Kachman |
331 | Training Language Models for Social Deduction with Multi-Agent Reinforcement Learning | Bidipta Sarkar, Warren Xia, Karen Liu, Dorsa Sadigh |
630 | TACTIC: Task-Agnostic Contrastive pre-Training for Inter-Agent Communication | Peihong Yu, Manav Mishra, Syed Zaidi, Pratap Tokekar |
1374 | MOSMAC: A Multi-agent Reinforcement Learning Benchmark on Sequential Multi-Objective Tasks | Minghong Geng, Shubham Pateria, Budhitama Subagdja, Ah-Hwee Tan |
481 | Compositional Shielding and Reinforcement Learning for Multi-Agent Systems | Asger Horn Brorholt, Kim Guldstrand Larsen, Christian Schilling |
795 | Decentralized Planning Using Probabilistic Hyperproperties | Francesco Pontiggia, Filip Macák, Roman Andriushchenko, Michele Chiari, Milan Ceska |
1338 | Towards Efficient Online Goal Recognition through Deep Learning | Lorenzo Serina, Mattia Chiari, Alfonso Gerevini, Luca Putelli, Ivan Serina |
427 | Translating Multi-Agent Modal Logics of Knowledge and Belief into Decidable First-Order Fragments | Qihui Feng, Hannah Wilk, Shakil M Khan, Gerhard Lakemeyer |
729 | Salience-Invariant Consistent Policy Learning for Generalization in Visual Reinforcement Learning | Jingbo Sun, Songjun Tu, Qichao Zhang, Ke Chen, Dongbin Zhao |
1038 | Robustness of Epistemic Gossip Protocols Against Data Loss | Yoshikatsu Kobayashi, Koji Hasebe |
37 | Impact Measures for Gradual Argumentation Semantics | Caren Al Anaissy, Jérôme Delobelle, Srdjan Vesic, Bruno Yun |
308 | Investigating the Perspective of Non-Native Speakers on Foreigner-Directed Speech using Virtual Agents: The Role of Racial Ingroup Affiliation and Language Proficiency on Perception and Comprehension | Ohenewa Bediako Akuffo, Birgit Lugrin |
544 | Beyond Words: Integrating Personality Traits and Context-Driven Gestures in Human-Robot Interactions | Tahsin Tariq Banna, Dr. Sejuti Rahman, Dr. Mohammad Tareq |
301 | Near-Linear Time Leader Election in Multiagent Networks | Ajay Kshemkalyani, Manish Kumar, Anisur Rahaman Molla, Gokarna Sharma |
696 | Housing Market on Networks | Xinwei Song, Tianyi Yang, Dengji Zhao |
303 | Curiosity-Driven Partner Selection Accelerates Convention Emergence in Language Games | Chin-Wing Leung, Paolo Turrini, Ann Nowe |
1218 | Hierarchical Learning-based Graph Partition for Large-scale Vehicle Routing Problems | Yuxin Pan, Ruohong Liu, Yize Chen, Zhiguang Cao, Fangzhen Lin |
244 | MAGNET: A Multi-Agent Graph Neural Network for Efficient Bipartite Task Assignment | Donald Loveland, James Usevitch, Zachary Serlin, Danai Koutra, Rajmonda S. Caceres |
398 | Incentives for Early Arrival in Cost Sharing | Junyu Zhang, Yao Zhang, Yaoxin Ge, Dengji Zhao, Hu Fu, Zhihao Gavin Tang, Pinyan Lu |
1265 | Conformal Set-based Human-AI Complementarity with Multiple Experts | Helbert Paat, Guohao Shen |
1026 | Uncertainty Expression for Human-Robot Task Communication | David Porfirio, Mark Roberts, Laura M. Hiatt |
162 | Tighter Value-Function Approximations for POMDPs | Merlijn Krale, Wietze Koops, Sebastian Junges, Thiago D. Simão, Nils Jansen |
534 | Synergistic Traffic Assignment | Thomas Bläsius, Adrian Feilhauer, Markus Jung, Moritz Laupichler, Peter Sanders, Michael Zündorf |
1389 | CAMP: Collaborative Attention Model with Profiles for Vehicle Routing Problems | Chuanbo Hua, Federico Berto, Jiwoo Son, Seunghyun Kang, Changhyun Kwon, Jinkyoo Park |
999 | Policy Graphs and Intention: answering ‘why’ and ‘how’ from a telic perspective | Victor Gimenez-Abalos, Sergio Alvarez-Napagao, Adrián Tormos, Ulises Cortés, Javier Vazquez-Salceda |
1273 | On Learning Informative Trajectory Embeddings for Imitation, Classification and Regression | Zichang Ge, Changyu Chen, Arunesh Sinha, Pradeep Varakantham |
67 | Online Preference-based Reinforcement Learning with Self-augmented Feedback from Large Language Model | Songjun Tu, Jingbo Sun, Qichao Zhang, Xiangyuan Lan, Dongbin Zhao |
865 | Greedy ABA Learning for Case-Based Reasoning | Emanuele De Angelis, Maurizio Proietti, Francesca Toni |
391 | Bayesian Collaborative Bandits with Thompson Sampling for Improved Outreach in Maternal Health | Arpan Dasgupta, Gagan Jain, Arun Suggala, Karthikeyan Shanmugam, Milind Tambe, Aparna Taneja |
1207 | Self-Interpretable Reinforcement Learning via Rule Ensembles | Yue Yang, Fan Yang, Yu Bai, Hao Wang |
905 | An Improved Mechanism for Pricing Ride-Hailing Fares | Marek Adamczyk, Maurycy Borkowski, Michał Pawłowski |
73 | To Spend or to Gain: Online Learning in Repeated Karma Auctions | Damien Berriaud, Ezzat Elokda, Devansh Jalota, Emilio Frazzoli, Marco Pavone, Florian Dorfler |
455 | Game-Theoretic Goal Recognition in Time-Sensitive Applications | Sara Bernardini, Fabio Fagnani, Santiago Franco |
518 | Alternating-time Temporal Logic with Stochastic Abilities | Gabriel Ballot, Vadim Malvone, Jean Leneutre, Jingxuan Ma, Mourad Leslous |
442 | Networked Agents in the Dark: Team Value Learning under Partial Observability | Guilherme S. Varela, Alberto Sardinha, Francisco S. Melo |
207 | Nucleolus Credit Assignment for Effective Coalitions in Multi-agent Reinforcement Learning | Yugu Li, Zehong Cao, Jianglin Qiao, Siyi Hu |
338 | Tackling Uncertainties in Multi-Agent Reinforcement Learning through Integration of Agent Termination Dynamics | Somnath Hazra, Pallab Dasgupta, Soumyajit Dey |
1377 | Byzantine Game Theory: Sun Tzu’s Boxes | Andrei Constantinescu, Roger Wattenhofer |
567 | Adaptive Bi-Level Multi-Robot Task Allocation and Learning under Uncertainty with Temporal Logic Constraints | Xiaoshan Lin, Roberto Tron |
837 | From Natural Language to Extensive-Form Game Representations | Shilong Deng, Yongzhao Wang, Rahul Savani |
827 | Resource Task Games | Jessica L. Newman, Enrico Gerding, Enrico Marchioni, Baharak Rastegari |
1064 | The Price of Anarchy in Spatial Social Choice | James Patrick Bailey, Craig Tovey |
646 | Loss of Plasticity: A New Perspective on Solving Multi-Agent Exploration for Sparse Reward Tasks | Zehua Zang, Chuxiong Sun, Lixiang Liu, Fuchun Sun, Changwen Zheng |
468 | Learning with Limited Shared Information in Multi-agent Multi-armed Bandit | Junning Shao, Siwei Wang, Zhixuan Fang |
1113 | Multi-Objective Planning with Contextual Lexicographic Reward Preferences | Pulkit Rustagi, Yashwanthi Anand, Sandhya Saisubramanian |
335 | Fairly Allocating Goods in Parallel | Rohan Garg, Alexandros Psomas |
394 | Harmonious Balanced Partitioning of a Network of Agents | Pulkit Agarwal, Harshvardhan Agarwal, Vaibhav Raj, Swaprava Nath |
1319 | Beyond Goal Recognition: A Reinforcement Learning-based Approach to Inferring Agent Behaviour | Sheryl Mantik, Michael Dann, Minyi Li, Huong Ha, Julie Porteous |
609 | Combining Planning and Reinforcement Learning for Solving Relational Multiagent Domains | Nikhilesh Prabhakar, Ranveer Singh, Harsha Kokel, Sriraam Natarajan, Prasad Tadepalli |
944 | Neural DNF-MT: A Neuro-symbolic Approach for Learning Interpretable and Editable Policies | Kexin Gu Baugh, Luke Dickens, Alessandra Russo |
845 | Ranking Joint Policies in Dynamic Games using Evolutionary Dynamics | Natalia Koliou, George Vouros |
1346 | Parameterized Algorithms for Multiagent Pathfinding on Trees | Argyrios Deligkas, Eduard Eiben, Robert Ganian, Iyad A. Kanj, Ramanujan Sridharan |
867 | OGS-SLAM: Hybrid ORB-Gaussian Splatting SLAM | Xiaohan Li, Wenxiang Shen, Dong Liu, Jun Wu |
738 | Monte Carlo Tree Search with Velocity Obstacles for safe and efficient motion planning in dynamic environments | Lorenzo Bonanni, Daniele Meli, Alberto Castellini, Alessandro Farinelli |
942 | Emergence of Recursive Language through Bootstrapping and Iterated Learning | Vikas Kumar, Ajin George Joseph |
111 | AdaCred: Adaptive Causal Decision Transformers with Feature Crediting | Hemant Kumawat, Saibal Mukhopadhyay |
907 | Computing Efficient and Envy-Free Allocations under Dichotomous Preferences using SAT | Ari Conati, Andreas Niskanen, Ronald De Haan, Matti Järvisalo |
1034 | Uncertain Machine Ethics Planning | Simon Kolker, Louise A. Dennis, Ramon Fraga Pereira, Mengwei Xu |
874 | Formalising Overdetermination in a Labelled Transition System | Gauvain Bourgne, Camilo Sarmiento, Jean Gabriel Gustave Ganascia |
63 | Multi-agent reinforcement learning in the all-or-nothing public goods game on networks | Benedikt Valentin Meylahn |
1068 | An AI-Driven Card Playing Robot: An Empirical Study on Communicative Style and Embodiment with Elderly Adults | Michael Banck, Elisabeth Ganal, Hanna-Finja Weichert, Frank Puppe, Birgit Lugrin |
129 | FLIGHT: Facility Location Integrating Generalized, Holistic Theory of Welfare | Avyukta Manjunatha Vummintala, Shivam Gupta, Shweta Jain, Sujit Gujar |
29 | Enhancing Graph-based Coordination with Evolutionary Algorithms for Episodic Multi-agent Reinforcement Learning | Kexing Peng, Pengyi Li, Jianye Hao |
533 | Dynamic Sight Range Selection in Multi-Agent Reinforcement Learning | Weichen Liao, Ti-Rong Wu, I-Chen Wu |
1102 | Indifferential Privacy: A New Paradigm and Its Applications to Optimal Matching in Dark Pool Auctions | Antigoni Polychroniadou, T-H. Hubert Chan, Adya Agrawal |
1226 | Full Proportional Justified Representation | Yusuf Hakan Kalayci, Jiasen Liu, David Kempe |
977 | LTL Verification of Memoryful Neural Agents | Mehran Hosseini, Alessio Lomuscio, Nicola Paoletti |
165 | Incentivizing Truth Exploration and Honest Reporting: A Contract Design Approach | Yuming Shao, Zhixuan Fang |
937 | Bidding Games on Markov Decision Processes with Quantitative Reachability Objectives | Guy Avni, Martin Kureƒçka, Kaushik Mallik, Petr Novotný, Suman Sadhukhan |
1009 | Agent-based Modeling and Simulation of Ambiguity in Catastrophe Insurance Markets | Yu Bi, Lingxiao Zhao, Jinyun Tong, Zhe Feng, Carmine Ventre |
1307 | Preventing Misinformation with Redundancy in Emergent Communication | Fábio Vital, Alberto Sardinha, Francisco S. Melo |
1049 | Sea-cret Agents: Maritime Abduction for Region Generation to Expose Dark Vessel Trajectories | Divyagna Bavikadi, Nathaniel Lee, Chad Parvis, Paulo Shakarian |
1005 | HAVA: Hybrid Approach to Value-Alignment through Reward Weighing for Reinforcement Learning | Kryspin Varys, Federico Cerutti, Adam Sobey, Timothy J. Norman |
662 | Global Behavior of Learning Dynamics in Zero-Sum Games with Memory Asymmetry | Yuma Fujimoto, Kaito Ariu, Kenshi Abe |
908 | Conditional Max-Sum for Asynchronous Multiagent Decision Making | Dimitrios Troullinos, Georgios Chalkiadakis, Ioannis Papamichail, Markos Papageorgiou |
462 | Opinion Dynamics with Median Aggregation | Petra Berenbrink, Martin Hoefer, Dominik Kaaser, Marten Maack, Malin Rau, Lisa Wilhelmi |
671 | Fair Allocation of Divisible Goods under Non-Linear Valuations | Haris Aziz, Zixu He, Xinhang Lu, Kaiyang Zhou |
797 | Games in Public Announcement: How to Reduce System Losses in Optimistic Blockchain Mechanisms | Siyuan Liu, Yulong Zeng |
224 | Factorised Active Inference for Strategic Multi-Agent Interactions | Jaime Ruiz-Serra, Patrick Sweeney, Michael Harre |
27 | The Strong Core of Housing Markets with Partial Order Preferences | Ildikó Schlotter, Lydia Mirabel Mendoza-Cadena |
511 | Simplifying imperfect recall games | Hugo Gimbert, Soumyajit Paul, B. Srivathsan |
628 | Revisiting Communication Efficiency in Multi-Agent Reinforcement Learning from the Dimensional Analysis Perspective | Chuxiong Sun, Peng He, Rui Wang, Changwen Zheng |
933 | Selecting Interlacing Committees | Chris Dong, Martin Bullinger, Tomasz Wƒös, Larry Birnbaum, Edith Elkind |
1250 | ShipNaviSim: Data-Driven Simulation for Real-World Maritime Navigation | Quang Anh Pham, Janaka Chathuranga Brahmanage, Akshat Kumar |
571 | Socratic: Enhancing Human Teamwork via AI-enabled Coaching | Sangwon Seo, Bing Han, Rayan Ebnali Harari, Roger Daglius Dias, Marco A. Zenati, Eduardo Salas, Vaibhav V. Unhelkar |
418 | Robust Policy Learning for Multi-UAV Collision Avoidance with Causal Feature Selection | Jiafan Zhuang, Gaofei Han, Zihaoxia, Che Lin, Boxi Wang, Wenji Li, Wangdongliang, Zhun Fan, Ruichu Cai, Zhifeng Hao |
1293 | Maximizing Value in Challenge the Champ Tournaments | Umang Bhaskar, Juhi Chaudhary, Palash Dey |
470 | The Bakers and Millers Game with Restricted Locations | Simon Krogmann, Pascal Lenzner, Alexander Skopalik |
756 | An Extended Benchmarking of Multi-Agent Reinforcement Learning Algorithms in Complex Fully Cooperative Tasks | George Papadopoulos, Andreas Kontogiannis, Foteini Papadopoulou, Chaido Poulianou, Ioannis Koumentis, George Vouros |
709 | FedRLHF: A Convergence-Guaranteed Federated Framework for Privacy-Preserving and Personalized RLHF | Flint Xiaofeng Fan, Cheston Tan, Yew-Soon Ong, Roger Wattenhofer, Wei Tsang Ooi |
245 | EduQate: Generating Adaptive Curricula through RMABs in Education Settings | Sidney Tio, Dexun Li, Pradeep Varakantham |
360 | The Metric Distortion of Randomized Social Choice Functions: C1 Maximal Lottery Rules and Simulations | Fabian Frank, Patrick Lederer |
472 | Fair Division in a Variable Setting | Harish Chandramouleeswaran, Prajakta Nimbhorkar, Nidhi Rathi |
24 | Order Symmetry: A New Fairness Criterion for Assignment Mechanisms | Rupert Freeman, Geoffrey Pritchard, Mark C. Wilson |
580 | Planning, scheduling, and execution on the Moon: the CADRE technology demonstration mission | Gregg Rabideau, Joseph A. Russino, Andrew Branch, Nihal N. Dhamani, Tiago Vaquero, Steve Chien, Jean-Pierre De La Croix, Federico Rossi |
1289 | Optimising expectation with guarantees for window mean payoff in Markov decision processes | Pranshu Gaba, Shibashis Guha |
441 | A Minimax-Bayes Approach to Ad Hoc Teamwork | Victor Villin, Thomas Kleine Buening, Christos Dimitrakakis |
714 | Single-Agent Planning in a Multi-Agent System: A Unified Framework for Type-Based Planners | Fengming Zhu, Fangzhen Lin |
1103 | Dynamic Coalition Structure Detection in Natural-Language-based Interactions | Abhishek Ninad Kulkarni, Andy Liu, Jean-Raphaël Gaglione, Daniel Fried, Ufuk Topcu |
530 | Rational Capability in Concurrent Games | Yinfeng Li, Emiliano Lorini, Munyque Mittelmann |
375 | FGLight: Learning neighbor-level information for Traffic Signal Control | Hang Xiao, Huale Li, Shuhan Qi, Jiajia Zhang, Dingzhong Cai |
170 | Learning Graph Representation of Agent Diffusers | Youcef Djenouri, Nassim Belmecheri, Tomasz Pawel Michalak, Jan Dubiński, Ahmed Nabil Belbachir, Anis Yazidi |
1283 | Emit As You Go: Enumerating Edges of a Spanning Tree | Katrin Casel, Stefan Neubert |
526 | Unveiling Decision Intention for Cooperative Multi-Agent Reinforcement Learning | Zeren Zhang, Zhiwei Xu, Guangchong Zhou, Dapeng Li, Bin Zhang, Guoliang Fan |
1140 | Leveraging Large Language Models for Effective and Explainable Multi-Agent Credit Assignment | Kartik Nagpal, Dayi Ethan Dong, Negar Mehr |
265 | Algorithmically Fair Maximization of Multiple Submodular Objective Functions | Georgios Amanatidis, Georgios Birmpas, Philip Lazos, Stefano Leonardi, Rebecca Reiffenhäuser |
704 | Consistency Policy with Categorical Critic for Autonomous Driving | Xing Fang, Qichao Zhang, Haoran Li, Dongbin Zhao |
587 | ApproxED: Approximate Exploitability Descent via Learned Best Responses | Carlos Martin, Tuomas Sandholm |
787 | PMAT: Optimizing Action Generation Order in Multi-Agent Reinforcement Learning | Kun Hu, Muning Wen, Xihuai Wang, Shao Zhang, Yiwei Shi, Minne Li, Minglong Li, Ying Wen |
762 | Mitigating Value Conflicts with Computational Theory of Mind | Emre Erdogan, Hüseyin Aydın, Frank Dignum, Rineke Verbrugge, Pinar Yolum |
657 | More Efficient Sybil Detection Mechanisms Leveraging Resistance of Users to Attack Requests | Ali Safarpoor Dehkordi, Ahad N. Zehmakan |
813 | Temporal Network Creation Games: The Impact of Non-Locality and Terminals | Davide Bilò, Sarel Cohen, Tobias Friedrich, Hans Gawendowicz, Nicolas Klodt, Pascal Lenzner, George Skretas |
279 | Higher-Order Belief in Incomplete Information MAIDs | Francis Rhys Ward, Jack Foxabbott, Rohan Subramani |
950 | Automating Curriculum Learning for Reinforcement Learning using a Skill-Based Bayesian Network | Vincent Hsiao, Mark Roberts, Laura M. Hiatt, George Konidaris, Dana S. Nau |
1189 | Counterfactual Explanations for Model Ensembles Using Entropic Risk Measures | Erfaun Noorani, Pasan Dissanayake, Faisal Hamman, Sanghamitra Dutta |
457 | Artificial Agents Mitigate The Punishment Dilemma Of Indirect Reciprocity | Alexandre S. Pires, Fernando P. Santos |
495 | On the Hardness of Fair Allocation under Ternary Valuations | Zack Fitzsimmons, Vignesh Viswanathan, Yair Zick |
500 | Tackling Temporal Deontic Challenges with Equilibrium Logic | Davide Soldà, Pedro Cabalar, Agata Ciabattoni, Emery A. Neufeld |
277 | FORM: Learning Expressive and Transferable First-Order Logic Reward Machines | Leo Ardon, Daniel Furelos-Blanco, Roko Parać, Alessandra Russo |
730 | Equilibrium Analysis in Markets with Asymmetric Utility Functions | Martin Bichler, Markus Ewert, Axel Ockenfels |
1008 | Extending Consensus-based Task Allocation Algorithms with Bid Intercession to Foster Mixed-Initiative | Victor Guillet, Charles Lesire, Gauthier Picard, Christophe Grand |
420 | Fairness and Optimality in Routing | Sreenivas Gollapudi, Kostas Kollias, Alkmini Sgouritsa, Ali Kemal Sinop |
297 | Simulating and Evaluating Generative Modeling and Collaborative Filtering in Complex Social Networks | Wen Dong, Fairul Mohd-Zaid |
765 | Free Argumentative Exchanges for Explaining Image Classifiers | Avinash Kori, Antonio Rago, Francesca Toni |
712 | Tackling Sparsity in Designated Driver Dispatch with Multi-Agent Reinforcement Learning | Jiaxuan Jiang, Ling Pan, Lin Zhou, Longbo Huang, Zhixuan Fang |
579 | Condorcet Winners and Anscombe’s Paradox Under Weighted Binary Voting | Carmel Baharav, Andrei Constantinescu, Roger Wattenhofer |
146 | On the Fairness of Additive Welfarist Rules | Karen Frilya Celine, Warut Suksompong, Sheung Man Yuen |
1233 | Reinforcement Learning Based Simulated Annealing | Nathan Qiu, Daniel Dali Liang |
1062 | On the Gale-Shapley Algorithm for Stable Matchings with a Partial Honesty Nash Refinement | James Patrick Bailey, Craig Tovey |
89 | Learning in Games with Progressive Hiding | Benjamin Heymann, Marc Lanctot |
607 | Bottom-Up Reputation Promotes Cooperation with Multi-Agent Reinforcement Learning | Tianyu Ren, Xuan Yao, Yang Li, Xiao-Jun Zeng |
167 | Personality-Driven Decision Making in LLM-Based Autonomous Agents | Lewis Newsham, Daniel Prince |
243 | Human-Agent Coordination in Games under Incomplete Information via Multi-Step Intent | Shenghui Chen, Ruihan Zhao, Sandeep P. Chinchali, Ufuk Topcu |
664 | EFX Allocations and Orientations on Bipartite Multi-graphs: A Complete Picture | Mahyar Afshinmehr, Alireza Danaei, Mehrafarin Kazemi, Kurt Mehlhorn, Nidhi Rathi |
4 | Logic of Knowledge and Cognitive Ability | Jia Tao, Xinran Zhang |
968 | Certified Guidance for Planning with Deep Generative Models | Francesca Cairoli, Francesco Giacomarra, Mehran Hosseini, Nicola Paoletti |
778 | GUIDE-CoT: Goal-driven and User-Informed Dynamic Estimation for Pedestrian Trajectory using Chain-of-Thought | Sungsik Kim, Baek Janghyun, Jinkyu Kim, Jaekoo Lee |
798 | Predictability Awareness for Efficient and Robust Multi-Agent Coordination | Román Chiva Gil, Daniel Jarne Ornia, Khaled A. Mustafa, Javier Alonso-Mora |
119 | On the limits of agency in agent-based models | Ayush Chopra, Shashank Kumar, Nurullah Giray Kuru, Ramesh Raskar, Arnau Quera-Bofarull |
6 | Enhancing Offline Reinforcement Learning with Curriculum Learning-Based Trajectory Valuation | Amir Abolfazli, Zekun Song, Avishek Anand, Wolfgang Nejdl |
819 | Multi-Ship Future Interaction Trajectory Prediction via Pre-Initializer Diffusion Model | Kun Ma, Qilong Han, Jingzheng Yao |
74 | Offline Goal-Conditioned Reinforcement Learning with Elastic-Subgoal Diffused Policy Learning | Yaocheng Zhang, Yuanheng Zhu, Yuqian Fu, Songjun Tu, Dongbin Zhao |
621 | Who Am I Dealing With? Explaining the Designer’s Hidden Intentions | Turgay Caglar, Sarath Sreedharan, Mor Vered |
1255 | Mean Field Correlated Imitation Learning | Zhiyu Zhao, Chengdong Ma, Qirui Mi, Ning Yang, Xue Yan, Mengyue Yang, Haifeng Zhang, Jun Wang, Yaodong Yang, |
915 | On Diffusion Models for Multi-Agent Partial Observability: Shared Attractors, Error Bounds, and Composite Flow | Tonghan Wang, Heng Dong, Yanchen Jiang, David C. Parkes, Milind Tambe |
154 | Boosting Sortition via Proportional Representation | Soroush Ebadian, Evi Micha |
930 | Taming Multi-Agent Reinforcement Learning with Estimator Variance Reduction | Taher Jafferjee, Juliusz Ziomek, Tianpei Yang, Zipeng Dai, Jianhong Wang, Matthew E. Taylor, Kun Shao, Jun Wang, David Henry Mguni, |
592 | SCMRAG: Self-Corrective Multihop Retrieval Augmented Generation System for LLM Agents | Rishabh Agrawal, Murtaza Asrani, Hadi Youssef, Apurva Narayan |
564 | Maximizing Truth Learning in a Social Network is NP-hard | Filip Úradník, Amanda Wang, Jie Gao |
643 | Value Iteration for Learning Concurrently Executable Robotic Control Tasks | Sheikh A. Tahmid, Gennaro Notomista |
357 | Learning Symbolic Task Decompositions for Multi-Agent Teams | Ameesh Shah, Niklas Lauffer, Thomas Chen, Nikhil Pitta, Sanjit A. Seshia |
523 | Temporal Fair Division of Indivisible Items | Edith Elkind, Alexander Lam, Mohamad Latifian, Tzeh Yuan Neoh, Nicholas Teh |
667 | Human-Aligned Skill Discovery: Balancing Behaviour Exploration and Alignment | Maxence Hussonnois, Thommen George Karimpanal, Santu Rana |
337 | Towards Fair and Efficient Public Transportation: A Bus Stop Model | Martin Bullinger, Edith Elkind, Mohamad Latifian |
900 | Agent-Based Analysis of Green Disclosure Policies and Their Market-Wide Impact on Firm Behavior | Lingxiao Zhao, Maria Polukarov, Carmine Ventre |
1022 | Generalised BDI Planning | Felipe Meneguzzi, Ramon Fraga Pereira, Nir Oren |
887 | A Simple Integration of Epistemic Logic and Reinforcement Learning | Thorsten Engesser, Thibaut Le Marre, Emiliano Lorini, François Schwarzentruber, Bruno Zanuttini |
1304 | Surprise! Surprise! Learn and Adapt | Huma Samin, Dylan J. Walton, Nelly Bencomo |
1036 | xSRL: Safety-Aware Explainable RL – Safety as a Product of Explainability | Risal Shahriar Shefin, Md Asifur Rahman, Thai Le, Sarra Alqahtani |
963 | Local Topological Information as a Powerful Enhancer for Generalizable Neural Method in Travelling Salesman Problem | Xiaoxin Bai, Junyang Yang, Shengchao Yuan, Yinghao Zhang, Hanqian Wu |
917 | Policy Abstraction and Nash Refinement in Tree-Exploiting PSRO | Christine Konicki, Mithun Chakraborty, Michael P. Wellman |
1303 | DUPRE: Data Utility Prediction for Efficient Data Valuation | Pham Kieu Thao Nguyen, Rachael Hwee Ling Sim, Quoc Phong Nguyen, See-Kiong Ng, Bryan Kian Hsiang Low |
803 | Improving Policy Optimization via 𝜺-Retrain | Luca Marzari, Priya L. Donti, Changliu Liu, Enrico Marchesini |
943 | Voter Model Meets Rumour Spreading: A Study of Consensus Protocols on Graphs with Agnostic Nodes | Marcelo Matheus Gauy, Anna Abramishvili, Eduardo Colli, Tiago Madeira, Frederik Mallmann-Trenn, Vinícius Franco Vasconcelos, David Kohan Marzagao |
1153 | On Some Fundamental Problems for Multi-Agent Systems Over Multilayer Networks | Daniel Rosenkrantz, Madhav Marathe, Zirou Qiu, S. S. Ravi, Richard Stearns |
112 | Candidate nomination for Condorcet-consistent voting rules | Ildikó Schlotter, Katarína Cechlárová |
739 | Why Instant-Runoff Voting Is So Resilient to Coalitional Manipulation: Phase transitions in the Perturbed Culture | François Durand |
14 | Automatic Verification of Linear Integer Planning Programs via Forgetting in LIAUPF | Liangda Fang, Shikang Chen, Xiaoman Wang, Xiaoyou Lin, Chenyi Zhang, Qingliang Chen, Quanlong Guan, Kaile Su |
436 | Model and Mechanisms of Consent for Responsible Autonomy | Anastasia S. Apeiron, Davide Dell’Anna, Pradeep K. Murukannaiah, Pinar Yolum |
1020 | Feature Engineering for Agents: An Adaptive Cognitive Architecture for Interpretable ML Monitoring | Gusseppe Bravo-Rocca, Peini Liu, Jordi Guitart, Rodrigo M Carrillo-Larco, Ajay Dholakia, David Ellison |
870 | Selfish Behavior and Resource Competition in Multi-Agent Systems | Costas Courcoubetis, Antonis Dimakis |
1208 | 𝛽-DQN: Improving Deep Q-Learning By Evolving the Behavior | Hongming Zhang, Fengshuo Bai, Chenjun Xiao, Chao Gao, Bo Xu, Martin Müller |
160 | EconoJax: A Fast & Scalable Economic Simulation in JAX | Koen Ponse, Aske Plaat, Niki Van Stein, Thomas M. Moerland |
263 | On the Effective Horizon of Inverse Reinforcement Learning | Yiqing Xu, Finale Doshi-Velez, David Hsu |
389 | Offline-to-Online Multi-Agent Reinforcement Learning with Offline Value Function Memory and Sequential Exploration | Hai Zhong, Xun Wang, Zhuoran Li, Longbo Huang |
272 | Real-World Testing Matters in Reinforcement Learning for Education | Anna Riedmann, Carlo D’Eramo, Birgit Lugrin |
805 | On Stateful Value Factorization in Multi-Agent Reinforcement Learning | Enrico Marchesini, Andrea Baisero, Rupali Bhati, Christopher Amato |
1317 | Ready, Bid, Go! On-Demand Delivery Using Fleets of Drones with Unknown, Heterogeneous Energy Storage Constraints | Mohamed S. Talamali, Genki Miyauchi, Thomas Watteyne, Micael Santos Couceiro, Roderich Gross |
282 | Robin Hood Reachability Bidding Games | Shaull Almagor, Guy Avni, Neta Dafni |
307 | Fast UCB-type algorithms for stochastic bandits with heavy and super heavy symmetric noise | Yuriy Dorn, Aleksandr Katrutsa, Ilgam Latypov, Pudovikov Andrey |
434 | Truthful mechanisms for linear bandit games with private contexts | Yiting Hu, Lingjie Duan |
159 | Dual Ensembled Multiagent Q-Learning with Hypernet Regularizer | Yaodong Yang, Guangyong Chen, Hongyao Tang, Furui Liu, Danruo Deng, Pheng-Ann Heng |
330 | Computing Efficient Envy-Free Partial Allocations of Indivisible Goods | Robert Bredereck, Andrzej Kaczmarczyk, Junjie Luo, Bin Sun |
169 | Causes and Strategies in Multiagent Systems | Sylvia S. Kerkhove, Natasha Alechina, Mehdi Dastani |
1139 | Discovery and Deployment of Emergent Robot Swarm Behaviors via Representation Learning and Real2Sim2Real Transfer | Connor Mattson, Varun Raveendra, Ricardo Vega, Cameron Nowzari, Daniel S. Drew, Daniel S. Brown |
941 | EnEnv 1.0: Energy Grid Environment for Multi-Agent Reinforcement Learning Benchmarking | Dominik Jacek Bogucki, Łukasz Eugeniusz Lepak, Sonam Parashar, Bart Blachowski, Paweł Wawrzyński |
116 | Adaptive Episode Length Adjustment for Multi-agent Reinforcement Learning | Byunghyun Yoo, Younghwan Shin, Hyunwoo Kim, Euisok Chung, Jeongmin Yang |
909 | Probabilistic Timed ATL | Wojciech Jamroga, Marta Kwiatkowska, Wojciech Penczek, Laure Petrucci, Teofil Sidoruk |
602 | k-Approval Veto: A Spectrum of Voting Rules Balancing Metric Distortion and Minority Protection | Fatih Erdem Kizilkaya, David Kempe |
929 | Practical Abstractions for Model Checking Continuous-Time Multi-Agent Systems | Yan Kim, Wojciech Jamroga, Wojciech Penczek, Laure Petrucci |
566 | Efficient and Optimal Policy Gradient Algorithm for Corrupted Multi-armed Bandits | Jiayuan Liu, Siwei Wang, Zhixuan Fang |
201 | Multi-agent Multi-armed Bandits with Minimum Reward Guarantee Fairness | Piyushi Manupriya, Himanshu, Sakethanath Jagarlapudi, Ganesh Ghalme |
91 | On the Power of Temporal Locality on Online Routing Problems | Swapnil Guragain, Gokarna Sharma |
424 | Insights Regarding the Success of Damping in Improving Belief Propagation | Uriel Zaed, Roie Zivan, Omer Lev |
1370 | Offline Multi-Agent Preference-based Reinforcement Learning with Agent-aware Direct Preference Optimization | Qian Kou, Mingyang Li, Zeyang Liu, Long Qian, Zhuoran Chen, Lipeng Wan, Xingyu Chen, Xuguang Lan |
946 | Multi-objective Reinforcement Learning with Nonlinear Preferences: Provable Approximation for Maximizing Expected Scalarized Return | Nianli Peng, Muhang Tian, Brandon Fain |
513 | Non-obvious Manipulability in Hedonic Games with Friends Appreciation Preferences | Michele Flammini, Maria Fomenko, Giovanna Varricchio |
1188 | Learning Collusion in Episodic, Inventory-Constrained Markets | Paul Friedrich, Barna Pásztor, Giorgia Ramponi |
334 | Welfare Approximation in Additively Separable Hedonic Games | Martin Bullinger, Vaggos Chatziafratis, Parnian Shahkar |
200 | Anytime Fairness Guarantees in Stochastic Combinatorial MABs: A Novel Learning Framework | Subham Pokhriyal, Shweta Jain, Ganesh Ghalme, Vaneet Aggarwal |
262 | Imitation from Diverse Behaviors: Wasserstein Quality Diversity Imitation Learning with Single-Step Archive Exploration | Xingrui Yu, Zhenglin Wan, David Mark Bossens, Yueming Lyu, Qing Guo, Ivor Tsang |
1029 | Smooth Information Gathering in Two-Player Noncooperative Games | Fernando Palafox, Jesse Milzman, David Fridovich-Keil, Dong Ho Lee, Ryan Park |
939 | Eliminating Majority Illusion | Foivos Fioravantes, Abhiruk Lahiri, Antonio Lauerbach, Lluís Sabater, Marie Diana Sieper, Samuel Wolf |
1347 | Enhancing Sub-Optimal Trajectory Stitching: Spatial Composition RvS for Offline RL | Sheng Zang, Zhiguang Cao, Bo An, Senthilnath Jayavelu, Xiaoli Li |
393 | Truthful and Welfare-maximizing Resource Scheduling with Application to Electric Vehicles | Ramsundar Anandanarayanan, Swaprava Nath, Prasant Misra |