TREC 2018 Core Track
- Collection available to participants: Now available at http://trec.nist.gov/data/wapost/
- Topics available to participants: April
- Runs due from participants: 21 August 2018
- TREC 2018 notebook paper deadline: mid-October
- TREC 2018 conference: TBD
The primary goals of the proposed core track are three-fold: (a) to bring together the community in a common track that could lead to a diverse set of participating runs, (b) to build one or more new test collections using more recently created documents, and (c) to establish a (new) test collection construction methodology that avoids the pitfalls of depth-k pooling.
As a side goal the track intends to:
- study the shortcomings of test collections constructed in the past
- experiment with new ideas for constructing test collections
- expand existing test collections in a number of dimensions:
- new participant tasks (ad-hoc/interactive)
- new relevance judgments (binary/multilevel)
- new pooling methods
- new assessment resources (NIST / Crowdsourcing)
- new retrieval systems contributing documents (Manual/Neural/Strong baselines)
The participants task will be ad-hoc search. Automatic and manual runs are encouraged. The organizers of the track will provide participants with a task (title/description/narrative of TREC topics) and allow participating sites to run the experiment as they wish as long as they contribute a ranked list of documents as an output.
DATA AND RESOURCES
Collection: TREC Washington Post Corpus The TREC Washington Post Corpus contains 608,180 news articles and blog posts from January 2012 through August 2017. The articles are stored in JSON format. Information about obtaining the collection can be found at http://trec.nist.gov/data/wapost/
Topics: Two sets of topics will be used: (a) 25 topics from the 2017 Common Core track and (b) 25 new topics to be developed by the NIST assessors.
Relevance Assessments: NIST assessors (50 topics)
System data structures (such as dictionaries, indices, thesauri, etc. whether constructed by hand or automatically) can be built using existing documents, topics, and relevance judgments, but these structures may not be modified in response to the new test topics. For example, you can’t add topic words that are not in your dictionary, nor may the system data structures be based in any way on the results of retrieving documents for the test topics and having a human look at the retrieved documents. A corollary of this rule is that your system may not be tuned to the TREC 2017 topics.
There are many possible methods for converting the supplied topic into queries that your system can execute. TREC defines two broad categories of methods, “automatic” and “manual”, based on whether manual intervention is used. Automatic construction is when there is no human involvement of any sort in the query construction process; manual construction is everything else. Note that this is a very broad definition of manual construction, including both runs in which the queries are constructed manually and then run without looking at the results, and runs in which the results are used to alter the queries using some manual operation.
The result of a run is generally a ranking of the top 10,000 documents retrieved for each topic. You may submit fewer than 1000 documents for a topic, but the ranked retrieval evaluation measures used in TREC evaluate to 10,000 and count empty ranks as irrelevant (so your score cannot be hurt by returning 10,000 documents). Similarly, systems that do not rank documents perform poorly as evaluated by these measures. Note that trec_eval evaluates a run based strictly on scores, not on the ranks you assign in your submission. If you want the precise ranking you submit to be evaluated, the scores must reflect it.
The format to use when submitting ranked results is as follows, using a space as the delimiter between columns. The width of the columns in the format is not important, but it is important to include all columns and have some amount of white space between the columns.
30 Q0 ZF08-175-870 0 4238 prise1 30 Q0 ZF08-306-044 1 4223 prise1 30 Q0 ZF09-477-757 2 4207 prise1 30 Q0 ZF08-312-422 3 4194 prise1 30 Q0 ZF08-013-262 4 4189 prise1 etc.
- the first column is the topic number.
- the second column is the query number within that topic. This is currently unused and should always be Q0.
- the third column is the official document number of the retrieved document and is the number found in the “docno” field of the document.
- the fourth column is the rank the document is retrieved,
- the fifth column shows the score (integer or floating point) that generated the ranking. This score must be in descending (non-increasing) order and is important to include so that we can handle tied scores (for a given run) in a uniform fashion (trec_eval sorts documents by these scores, not your ranks).
- the sixth column is called the “run tag” and should be an unique identifier for your group AND for the method used. That is, each run should have a different tag that identifies the group and the method that produced the run.
NIST will release a routine that checks for common errors in the result files including duplicate document numbers for the same topic nvalid document numbers, wrong format, and duplicate tags across runs. This routine will be made available to participants to check their runs for errors prior to submitting them. Submitting runs is an automatic process done through a web form, and runs that contain errors cannot be processed.
ASSESSMENT AND EVALUATION
Assessment: The pooling method is TBD.
Evaluation: Participating runs will be evaluated in terms of (a) their ability to rank relevant documents at the top of the returned list, and (b) their ability to contribute unique relevant documents to the pool.
- Evangelos Kanoulas (firstname.lastname@example.org), University of Amsterdam
- James Allan (email@example.com), University of Massachusetts
- Donna Harman (firstname.lastname@example.org), NIST