AISOC Team

Chair of Artificial Intelligence and Society · Ruhr University Bochum · UA Ruhr RC Trust

Group Photo

About

The AI and Society (AISOC) team is is headed by Prof. Bilal Zafar. We are part of the Faculty of Computer Science at Ruhr University Bochum and affiliated with the Research Center for Trustworthy Data Science and Security (RC Trust), and the Cluster of Excellence CASA.

Our work aims to advance the development of trustworthy Artificial Intelligence (AI) and Machine Learning (ML) systems. By integrating multi-disciplinary perspectives, we aim to establish a comprehensive and practical framework for AI/ML trustworthiness. The key questions guiding our current research are:

  • Quantifying LLM’s ability to understand the world: The performance of LLMs regularly matches or even exceeds humans on popular benchmarks. However, proficiency in solving benchmarks does not equate human-level reasoning. For instance, a high score on a graduate-level math benchmark rarely implies the same level of conceptual understanding as a graduate-level math student. Since LLMs understand the world very differently than humans, our work aims to quantify these discrepancies.
  • Measuring and calibrating trust in AI: Despite their impressive capabilities, AI models continue to exhibit vulnerabilities like bias, lack of faithful explanations and hallucinations. At the same time, it is not always clear what it means for a LLM to be biased and for an explanation to be satisfactory. Our work draws insights from social sciences to operationalize these concepts and mitigate problematic behaviors.
  • Deployment challenges: Current day AI models are so compute-intensive that efficient deployment requires specially designed acceleration procedures like weight quantization and token pruning. These strategies, while enhancing efficiency, can have an adverse impact on the predictive performance. Our work aims to understand the impact of acceleration procedures and balance efficiency with performance.

Latest News

Older News

Jan 2026 Work on Memories in ChatGPT, jointly led by Abhishek, Soumi, Qinyuan and Elisabeth got accepted at WWW'26.
Jan 2026 Debtanu's summer internship work hallucination detection in low resource languages got accepted at EACL'26.
Jan 2026 Qinyuan's paper on work Generalizing over Memorized Data in LLMs got accepted at ICLR'26.
Jan 2026 Leon joined the team as a PhD student. Welcome Leon!
Dec 2025 Our team co-organized the workshop on Metacognition in Generative AI (co-located with EurIPS) with Marcel Binz, Hamed Hassani, Nastaran Okati and Isabel Valera.
Nov 2025 Elisabeth's work on generative search was covered by ArsTechnica, Tagesschau, and Heise. Read the paper here.

Recent Publications and Preprints

  1. Characterizing Web Search in The Age of Generative AI
    Elisabeth Kirsten, Jost Grosse Perdekamp, Mihir Upadhyay, Krishna P. Gummadi, and Muhammad Bilal Zafar
    2025
  2. WWW
    The Algorithmic Self-Portrait: Deconstructing Memory in ChatGPT
    Abhisek Dash, Soumi Das, Elisabeth Kirsten, Qinyuan Wu, Sai Keerthana Karnam, Krishna P. Gummadi, Thorsten Holz, Muhammad Bilal Zafar, and Savvas Zannettou
    In The Web Conference, 2026
  3. Do LLM hallucination detectors suffer from low-resource effect?
    Debtanu Datta, Mohan Kishore Chilukuri, Yash Kumar, Saptarshi Ghosh, and Muhammad Bilal Zafar
    In European Chapter of the Association for Computational Linguistics, 2026
  4. Rote Learning Considered Useful: Generalizing over Memorized Data in LLMs
    Qinyuan Wu, Soumi Das, Mahsa Amani, Bishwamittra Ghosh, Mohammad Aflah Khan, Krishna P. Gummadi, and Muhammad Bilal Zafar
    In International Conference on Learning Representations, 2026
  5. Can LLMs Explain Themselves Counterfactually?
    Zahra Dehghanighobadi, Asja Fischer, and Muhammad Bilal Zafar
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, 2025
  6. Position is Power: System Prompts as a Mechanism of Bias in Large Language Models (LLMs)
    Anna Neumann, Elisabeth Kirsten, Muhammad Bilal Zafar, and Jatinder Singh
    In Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency, 2025
  7. The Impact of Inference Acceleration on Bias of LLMs
    Elisabeth Kirsten, Ivan Habernal, Vedant Nanda, and Muhammad Bilal Zafar
    In Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics, 2025
  8. KDD
    On Early Detection of Hallucinations in Factual Question Answering
    Ben Snyder, Marius Moisescu, and Muhammad Bilal Zafar
    In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2024