Task Definition: Given a check-worthy claim on a specific topic and a set of text snippets extracted from potentially-relevant webpages, return a ranked list of evidence snippets for the claim. Evidence snippets are those snippets that are useful in verifying the given claim. This task will run in Arabic.
Evidence Snippet: An evidence snippet is a text snippet from a Webpage that constitutes evidence supporting or refuting the claim.
This task is evaluated as a ranking task. Ranked list per topic will be evaluated using ranking evaluation measures (MAP, P@5,10…,P@30). Official measure is P@10.
Each team can submit up to 2 manual and 4 automatic runs as follows:
Submit one separate results file per run. Evidence snippets per topic must be sorted by rank (from rank 1 till n). For each run, use the following format.
The results file should include a ranking of top 100 evidence snippets per claim (tweet). It must include one tab-separated line per tweet formatted as follows:
topicID  tweetID  rank  snippetID  score  runIDCT20-AR-05  1219151214690041857  1  CT20-AR-05-0003-001  0.77  teamXrun1CT20-AR-05  1219151214690041857  2  CT20-AR-05-0005-004  0.74  teamXrun1CT20-AR-05  1219151214690041857  3  CT20-AR-05-0036-002  0.68  teamXrun1…Where the score is a number indicating the usefulness of the snippet for fact-checking the tweet, the rank is the rank of the snippet according to its score, and the runID is a unique ID for one of the runs of the team.
Given a check-worthy claim on a specific topic and a set of potentially-relevant Web pages, predict the veracity of the claim: TRUE or FALSE. This task will run in Arabic.
The task is a classical binary classification task. Evaluation measures are standard: Precision, recall, and F1. Official measure is macro-averaged F1.
Each team can submit up to 2 manual and 4 automatic runs as follows:
Submit one separate results file per run. For each run, use the following format.
The results file should include one tab-separated line per claim formatted as follows:
topicID  tweetID  label  runIDCT20-AR-05  1218603003755798529  FALSE  teamXrun1CT20-AR-05  1219151214690041857  TRUE  teamXrun1CT20-AR-05  1217636592908689409  FALSE  teamXrun1...Where the label is one of: [TRUE, FALSE].
Your result file MUST contain predictions for all claims from the respective input file. Otherwise, the scorer will not score this result file.