All Stories

  1. A Multi-Armed Bandit Framework for Personalized Sustainability Goal Assignment
  2. AIDME: A Scalable, Interpretable Framework for AI-Aided Scoping Reviews
  3. Impersonating the Crowd: Evaluating LLMs' Ability to Replicate Human Judgment in Misinformation Assessment
  4. Preaching to the ChoIR: Lessons IR Should Share with AI
  5. PILs of Knowledge: A Synthetic Benchmark for Evaluating Question Answering Systems in Healthcare
  6. The Magnitude of Truth: On Using Magnitude Estimation for Truthfulness Assessment
  7. Efficiency and Effectiveness of LLM-Based Summarization of Evidence in Crowdsourced Fact-Checking
  8. Mapping and Influencing the Political Ideology of Large Language Models using Synthetic Personas
  9. Report on the 14th Italian Information Retrieval Workshop (IIR 2024)
  10. The Elusiveness of Detecting Political Bias in Language Models: The Impact of Question Wording
  11. Generative AI for Energy: Multi-Horizon Power Consumption Forecasting using Large Language Models
  12. Understanding the Barriers to Running Longitudinal Studies on Crowdsourcing Platforms
  13. Combining Large Language Models and Crowdsourcing for Hybrid Human-AI Misinformation Detection
  14. Data Bias Management
  15. How Many Crowd Workers Do I Need? On Statistical Power When Crowdsourcing Relevance Judgments
  16. Combining Human and Machine Confidence in Truthfulness Assessment
  17. Preferences on a Budget: Prioritizing Document Pairs when Crowdsourcing Relevance Judgments
  18. Crowd_Frame: A Simple and Complete Framework to Deploy Complex Crowdsourcing Tasks Off-the-Shelf