Information retrieval effectiveness evaluation is often carried out by means of test collections. Many works investigated possible sources of bias in such an approach. We propose a systematic approach to identify bias and its causes, and to remove it, thus enforcing fairness in effectiveness evaluation by means of test collections.
Enhancing Fact-Checking: From Crowdsourced Validation to Integration with Large Language Models
Kevin Roitero
;Michael Soprano
;David La Barbera;Eddy Maddalena;Stefano Mizzaro
2024-01-01
Abstract
Information retrieval effectiveness evaluation is often carried out by means of test collections. Many works investigated possible sources of bias in such an approach. We propose a systematic approach to identify bias and its causes, and to remove it, thus enforcing fairness in effectiveness evaluation by means of test collections.File in questo prodotto:
File | Dimensione | Formato | |
---|---|---|---|
paper13.pdf
accesso aperto
Tipologia:
Versione Editoriale (PDF)
Licenza:
Creative commons
Dimensione
913.15 kB
Formato
Adobe PDF
|
913.15 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.