I am a Full Professor in Explainable AI at Maastricht University in the Department of Advanced Computing Sciences (DACS) where I am the Director of Research. I am also a member of the Explainable AI (XAI) research theme, which is an interdisciplinary field of research focusing on providing explanations of AI-driven systems. This includes intrinsically explainable approaches to AI, as well as methods that provide explanations for decisions made by "black-box" machine learning models. The ability of AI-driven systems to explain their decisions in human-understandable terms improves the trust we place in these systems and helps address issues related to fair and unbiased decision-making. We bring together researchers working on XAI in a diverse range of application areas, including recommender systems, computational social science, causal inference for life sciences, affective computing, computer vision, knowledge representation/reasoning and machine learning. We have expertise on interactive interfaces for XAI, computational argumentation for XAI, and the evaluation of XAI systems.

cartoon

(Cartoon by Erwin Suvaal from CVIII ontwerpers.)

My research is funded by projects in the field of human-computer interaction in artificial advice-giving systems, such as recommender systems; specifically developing the state-of-the-art for automatically generated explanations (transparency) and explanation interfaces (recourse and control). These include projects a EU Marie-Curie ITN on Interactive Natural Language Technology for Explainable Artificial Intelligence. Currently, I am representing Maastricht university as a Co-Investigator in the ROBUST consortium to carry out long term (10-years) research into trustworthy artificial intelligence. I am also a co-lab director of the TAIM lab, working on trustworthy media, in collaboration with UvA and RTL. I regularly shape national (On NWO-ENW domain advisory committee Computer Science, the Informaticatafel) and international scientific research programs (e.g., steering boards, editorial board of journals, and as program chair of conferences), and actively organize and contribute to high-level strategic workshops relating to responsible data science, both in the Netherlands and internationally. I am a senior member of the ACM. I regularly examine PhD dissertations, and review both national (primarily NWO) and international grant proposals.



Relevant keywords: explanations, natural language generation, human-computer interaction, personalization (recommender systems), intelligent user interfaces, diversity, filter bubbles, responsible data analytics.


News:

2024

5th of February Our manuscript (w. Bart Knijnenburg and Martijn Willemsen) "Measuring the Benefit of Increased Transparency and Control in News Recommendation" has been accepted for publication in AI Magazine.
9th of January Accepted paper with Federico Cau at ECIR2024 (Information Retrieval for Good track). Title: Navigating the Thin Line: Examining User Behavior in Search to Detect Engagement and Backfire Effects.
9th of January NORMalize aims to promote normative thinking in the design of recommender/information retrieval systems and has been accepted as a tutorial at CHIIR2024, March 10-14 in Sheffield. We (with Lien Michiels, Johannes Kruse, Alain Starke, and Sanne Vrijenhoek) hope to see you there!

2023

13th of December Congratulations to Dr. Tim Draws on receiving his PhD cum laude.
9th of November Our journal paper: "Nudges to Mitigate Confirmation Bias during Web Search for Opinion Formation: Support vs. Manipulation", has been accepted for ACM Transactions on the Web (TWEB). Led by Alisa Rieger and with Tim Draws and Mariete Theune.
9th of November Looking forward to the 1st Dutch Belgian Workshop on Recommender systems. Come and learn about the interesting work PhD candidates Roan and Dina are doing in our group on explainable artificial intelligence! DBWRS2023
9th of November Congratulations to PhD candidate Adarsa Sivaprasad for her first accepted paper at the XAI^3 Workshop at ECAI 2023. Pre-print The paper is titled: Evaluation of Human-Understandability of Global Model Explanations using Decision Tree.
9th of November Francesco Barile is representing the XAI group at BNAIC'23 with our joint work ''The Effectiveness of Different Group Recommendation Strategies for Different Group Compositions''
7th of July Report accepted to SIGIR Forum: Bauer, C., Carterette, B., Ferro, N., Fuhr, N. et al. (2023). Report on the Dagstuhl Seminar on Frontiers of Information Access Experimentation for Research and Education. SIGIR Forum.
2nd of June Two full papers accepted at the 1st international XAI conference:
Title: A co-design study for multi-stakeholder job recommender system explanations
Authors: Roan Schellingerhout, Francesco Barile and Nava Tintarev
Title: Explaining Search Result Stances to Opinionated People
Authors: Zhangyi Wu, Tim Draws, Federico Cau, Francesco Barile, Alisa Rieger and Nava Tintarev

2nd of June Two journal papers to appear in the UMUAI special issue on group recommender systems. Title: Evaluating Explainable Social Choice-based Aggregation Strategies for Group Recommendation.
Authors: Francesco Barile, Tim Draws, Oana Inel, Alisa Rieger, Shabnam Najafian, Amir Ebrahimi Fard, Rishav Hada, and Nava Tintarev.
Authors: Shabnam Najafian, Geoff Musick, Bart Knijnenburg, and Nava Tintarev.
Title: "How do People Make Decisions in Disclosing Personal Information in Tourism Group Recommendations in Competitive versus Cooperative Conditions?"

26th of April Upcoming talks and events: Panelist on the opportunities and challenges of artificial intelligence in our society as part of the Philips Innovation Award 15th of May; Panelist on Open Science and AI at UM Open Science Festival on the 25th of May; Invited talk at the SIAS research group and the Civic AI Lab 26th of May; Keynote at IS-EUD June.

17th of February Journal paper (TIIS) accepted: "Effects of AI and Logic-Style Explanations on Users' Decisions under Different Levels of Uncertainty". Authors: Federico Cau, Hanna Hauptmann, Davide Spano, and Nava Tintarev

16th of January The long-term program ROBUST ``Trustworthy AI systems for sustainable growth'' has been supported by NWO in the new Long Term Program with 25 million Euros! Ph.D. positions for 17 labs will be advertised here soon, including the TAIM lab (UM, UvA, RTL) on trustworthy media.

16th of January Publications accepted to IUI and CHI!
Cau, F. M., Hauptmann, H., Spano, L. D., & Tintarev, N. (2023). Supporting High-Uncertainty Decisions through AI and Logic-Style Explanations. IUI.
BEST PAPER AWARD: Yurrita, M., Draws, T., Balayn, A., Murray-Rust, D., Tintarev, N., & Bozzon, A. (2023). Disentangling Fairness Perceptions in Algorithmic Decision-Making: the Effects of Explanations, Human Oversight, and Contestability. CHI.