All Stories

  1. Optimizing LLMs with Direct Preferences: A Data Efficiency Perspective
  2. Hate Speech Detection with Generalizable Target-aware Fairness
  3. Understanding the Barriers to Running Longitudinal Studies on Crowdsourcing Platforms
  4. Fairness without Sensitive Attributes via Knowledge Sharing
  5. How Good are LLMs in Generating Personalized Advertisements?
  6. Editorial: Special Issue on Human in the Loop Data Curation
  7. Who Determines What Is Relevant? Humans or AI? Why Not Both?
  8. On the Impact of Showing Evidence from Peers in Crowdsourced Truthfulness Assessments
  9. Data Bias Management
  10. Perspectives on Large Language Models for Relevance Judgment
  11. On the Impact of Data Quality on Image Classification Fairness
  12. How Many Crowd Workers Do I Need? On Statistical Power When Crowdsourcing Relevance Judgments
  13. Human-in-the-loop Regular Expression Extraction for Single Column Format Inconsistency
  14. The Community Notes Observatory: Can Crowdsourced Fact-Checking be Trusted in Practice?
  15. Report on the 1st Workshop on Human-in-the-Loop Data Curation (HIL-DC 2022) at CIKM 2022
  16. A Data-Driven Analysis of Behaviors in Data Curation Processes
  17. Combining Human and Machine Confidence in Truthfulness Assessment
  18. Using Computers to Fact-Check Text and Justify the Decision
  19. Socio-Economic Diversity in Human Annotations
  20. Preferences on a Budget: Prioritizing Document Pairs when Crowdsourcing Relevance Judgments
  21. Does Evidence from Peers Help Crowd Workers in Assessing Truthfulness?
  22. Effects of Technological Interventions for Self-regulation: A Control Experiment in Learnersourcing
  23. Hierarchical Clustering of Corals using Image Clustering
  24. An Analysis of the Australian Political Discourse in Sponsored Social Media Content
  25. On the State of Reporting in Crowdsourcing Experiments and a Checklist to Aid Current Practices
  26. Charting the Design and Analytics Agenda of Learnersourcing Systems
  27. Report on the first workshop on bias in automatic knowledge graph construction at AKBC 2020
  28. Modelling User Behavior Dynamics with Embeddings
  29. The COVID-19 Infodemic
  30. How to make crowd workers earn an hourly wage
  31. On Understanding Data Worker Interaction Behaviors
  32. Can The Crowd Identify Misinformation Objectively?
  33. Representation learning for entity type ranking
  34. Health Card Retrieval for Consumer Health Search
  35. On Transforming Relevance Scales
  36. Understanding Worker Moods and Reactions to Rejection in Crowdsourcing
  37. Quality Control Attack Schemes in Crowdsourcing
  38. Health Cards for Consumer Health Search
  39. Implicit Bias in Crowdsourced Knowledge Graphs
  40. Scalpel-CD: Leveraging Crowdsourcing and Deep Probabilistic Modeling for Debugging Noisy Training Data
  41. Deadline-Aware Fair Scheduling for Multi-Tenant Crowd-Powered Systems
  42. All Those Wasted Hours
  43. Novel insights into views towards H1N1 during the 2009 Pandemic: a thematic analysis of Twitter data
  44. Non-parametric Class Completeness Estimators for Collaborative Knowledge Graphs—The Case of Wikidata
  45. Semantic Interlinking
  46. The Impact of Task Abandonment in Crowdsourcing
  47. The Evolution of Power and Standard Wikidata Editors: Comparing Editing Behavior over Time to Predict Lifespan and Volume of Edits
  48. Can User Behaviour Sequences Reflect Perceived Novelty?
  49. Moral Panic through the Lens of Twitter
  50. Investigating User Perception of Gender Bias in Image Search
  51. On Fine-Grained Relevance Scales
  52. On the Volatility of Commercial Search Engines and its Impact on Information Retrieval Research
  53. Crowd Anatomy Beyond the Good and Bad: Behavioral Traces for Crowd Worker Modeling and Pre-selection
  54. Measuring the Effect of Public Health Campaigns on Twitter: The Case of World Autism Awareness Day
  55. Augmenting Intelligence with Humans-in-the-Loop (HumL@WWW2018) Chairs' Welcome & Organization
  56. Chapter 4: Using Twitter as a Data Source: An Overview of Ethical, Legal, and Methodological Challenges
  57. Understanding Engagement through Search Behaviour
  58. Considering Assessor Agreement in IR Evaluation
  59. Modus Operandi of Crowd Workers
  60. An Introduction to Hybrid Human-Machine Information Systems
  61. Towards building a standard dataset for Arabic keyphrase extraction evaluation
  62. Scheduling Human Intelligence Tasks in Multi-Tenant Crowd-Powered Systems
  63. Contextualized ranking of entity types based on knowledge graphs
  64. A Tutorial on Leveraging Knowledge Graphs for Web Search
  65. The Relationship Between User Perception and User Behaviour in Interactive Information Retrieval Evaluation
  66. Hybrid human–machine information systems: Challenges and opportunities
  67. Pooling-based continuous evaluation of information retrieval systems
  68. Human Beyond the Machine: Challenges and Opportunities of Microtask Crowdsourcing
  69. The Dynamics of Micro-Task Crowdsourcing
  70. Understanding Malicious Behavior in Crowdsourcing Platforms
  71. Correct Me If I'm Wrong
  72. B-hist: Entity-centric search over personal web browsing history
  73. Hippocampus
  74. Effective named entity recognition for idiosyncratic web collections
  75. Entity disambiguation in tweets leveraging user social profiles
  76. Large-scale linked data integration using probabilistic reasoning and crowdsourcing
  77. NoizCrowd: A Crowd-Based Data Gathering and Management System for Noise Level Data
  78. Ontology-Based Word Sense Disambiguation for Scientific Literature
  79. TRank: Ranking Entity Types Using the Web of Data
  80. The Bowlogna ontology: Fostering open curricula and agile knowledge bases for Europe's higher education landscape
  81. ZenCrowd
  82. BowlognaBench—Benchmarking RDF Analytics
  83. Combining inverted indices and structured search for ad-hoc object retrieval
  84. Predicting the Future Impact of News Events
  85. From people to entities
  86. Visual interfaces for stimulating exploratory search
  87. Report on INEX 2009
  88. Why finding entities in Wikipedia is difficult, sometimes
  89. Leveraging personal metadata for Desktop search: The Beagle++ system
  90. Dear search engine: what's your opinion about...?
  91. Entity summarization of news articles
  92. Exploiting click-through data for entity retrieval
  93. Overview of the INEX 2009 Entity Ranking Track
  94. Ranking Entities Using Web Search Query Logs
  95. TAER
  96. The missing links
  97. An Architecture for Finding Entities on the Web
  98. Report on INEX 2008
  99. A Vector Space Model for Ranking Entities and Its Application to Expert Search
  100. How to Trace and Revise Identities
  101. L3S at INEX 2008: Retrieving Entities Using Structured Information
  102. Overview of the INEX 2008 Entity Ranking Track
  103. A Model for Ranking Entities and Its Application to Wikipedia
  104. Social recommendations of content and metadata
  105. Leveraging semantic technologies for enterprise search
  106. A Classification of IR Effectiveness Metrics
  107. L3S at INEX 2007: Query Expansion for Entity Ranking Using a Highly Accurate Ontology
  108. Ranking Categories for Web Search
  109. Semantically Enhanced Entity Ranking