Bálint Gyevnár
Safeguarding scientific integrity in the age of AI co-scientists

Hi, I am Bálint, an incoming postdoc at CMU.
(My name is pronounced BAH-lint [baːlint])
My primary research area is the science of AI co-scientists: how they amplify existing biases in research production, create epistemic blind spots, and homogenise approaches across disciplines. I work with Atoosa Kasirzadeh and Nihar Shah as a member of the Institute for Complex Social Dynamics led by Kevin Zollman.
During my PhD, I worked on explainable multi-agent reinforcement learning, which I like to describe as the study of giving interacting AI agents the ability to explain themselves. I researched how we can explain complex emergent behaviour in multi-agent systems (MAS) via the use of counterfactual reasoning. I was supervised by Stefano Albrecht, Shay Cohen, and Chris Lucas.
news
Aug 10, 2025 | I spent a week visiting the Center for Humans and Machines led by Iyad Rahwan at the Max Planck Institute, Berlin. |
---|---|
Jul 30, 2025 | I attended the 2025 Human-aligned AI Summer School in Prague, oragnised by the Alignment of Complex Systems Research and the Center for Theoretical Study at Charles University. |
Jun 18, 2025 | I attended the 2025 Bridging Responsible AI Divides (BRAID) Gathering in Manchester. |
Jun 11, 2025 | I attended RLDM 2025, the Multi-disciplinary Conference on Reinforcement Learning and Decision Making, in Dublin, where I have presented a poster on Objective Metrics for Explainable RL paper. |
Jun 07, 2025 | I gave a talk and presented a poster at the 9th Center for Human-Compatible AI Workshop on “AI Safety for Everyone”. |
May 26, 2025 | New preprint paper titled: Integrating Counterfactual Simulations with Language Models for Explaining Multi-Agent Behaviour. |
Mar 21, 2025 | I am co-organising a workshop on “Evaluating Explainable AI and Complex Decision-Making” co-located with ECAI ‘25. Call for papers found here. |