Bálint Gyevnár

Safeguarding scientific integrity in the age of AI scientists

prof_pic.jpg
Postdoctoral Research Associate at Carnegie Mellon University

My research is concerned with artificial intelligence (AI) in science and the science in AI. I am interested in computational (e.g. algorithms) and empirical methods (e.g. cognitive experiments) of understanding how AI can be used to safely and sustainably enhance human scientific discovery. I am working with Atoosa Kasirzadeh and Nihar Shah at the intersection of machine learning, cognitive science, and the philosophy of science. During my PhD, I worked on explainable multi-agent reinforcement learning under the supervision of Stefano Albrecht, Shay Cohen, and Chris Lucas at the University of Edinburgh, Scotland.

I am currently most curious about three questions:

  • How do we protect scientific integrity in the age of agentic AI scientists? What are the methodological pitfalls of scientific AI systems, and how do we create controlled computational tools with statistical guarantees to prevent them?
  • How does the use of AI in science affect the way scientists think and act? What are the epistemological dangers of delegating scientific thinking to AI systems, and how does AI change the high-level conceptual and low-level methodological steps of science?
  • How do scientific communities change as a result of scientific AI deployment? What are the effects of AI-accelerated scientific discovery on research communities’ interests, and what are scientists’ beliefs and desires for AI in science?

If you are curious about any of the above topics, then don’t hesitate to reach out through the various channels at the bottom of this page! I am currently based in Pittsburgh, PA, USA.

(My name is pronounced BAH-lint [baːlint])

news

Dec 19, 2025 Our paper Integrating Counterfactual Simulations with Language Models for Explaining Multi-Agent Behaviour was accepted to AAMAS 2026. Reach out if you are going to Cyprus next May!
Dec 06, 2025 I attended FAR.AI Alignment Workshop in San Diego just before NeurIPS, presenting my poster titled We Need a Rigorous Metascience of Artificial Intelligence.
Nov 14, 2025 I attended the workshop and wrote a post about the Cognitive Science of Mathematical Understanding, co-organised by Tania Lombrozo and Akshay Venkatesh.
Sep 15, 2025 I have started as a postdoc at Carnegie Mellon University at the Institute of Complex Social Dynamics working with Atoosa Kasirzadeh and Nihar Shah.
Aug 10, 2025 I spent a week visiting the Center for Humans and Machines led by Iyad Rahwan at the Max Planck Institute, Berlin.
Jul 30, 2025 I attended the 2025 Human-aligned AI Summer School in Prague, oragnised by the Alignment of Complex Systems Research and the Center for Theoretical Study at Charles University.
Jun 18, 2025 I attended the 2025 Bridging Responsible AI Divides (BRAID) Gathering in Manchester.

latest posts

selected publications

  1. Integrating Counterfactual Simulations with Language Models for Explaining Multi-Agent Behaviour
    In Proceedings of the 25th International Conference on Autonomous Agents and Multiagent Systems, Paphos, Cyprus, May 2026
  2. AI Safety for Everyone
    Balint Gyevnar*, and Atoosa Kasirzadeh*
    Nature Machine Intelligence, Apr 2025
  3. CHI
    People Attribute Purpose to Autonomous Vehicles When Explaining Their Behavior: Insights from Cognitive Science for Explainable AI
    In Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, Apr 2025
  4. Causal Explanations for Sequential Decision-Making in Multi-Agent Systems
    In Proceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems, Auckland, New Zealand, May 2024
  5. Bridging the Transparency Gap: What Can Explainable AI Learn From the AI Act?
    Balint Gyevnar, Nick Ferguson, and Burkhard Schafer
    In 26th European Conference on Artificial Intelligence, Sep 2023