Actions
shareshare link cite add Please grant OpenAIRE to access and update your ORCID works.This Research product is the result of merged Research products in OpenAIRE.
You have already added 0 works in your ORCID record related to the merged Research product.
See an issue? Give us feedback
Please grant OpenAIRE to access and update your ORCID works.
This Research product is the result of merged Research products in OpenAIRE.
You have already added 0 works in your ORCID record related to the merged Research product.
You have already added 0 works in your ORCID record related to the merged Research product.
Research data . Dataset . 2010
Webis Query Segmentation Corpus 2010 (Webis-QSeC-10)
Hagen, Matthias; Potthast, Martin; Stein, Benno; Bräutigam, Christof; Beyer, Anna;
Hagen, Matthias; Potthast, Martin; Stein, Benno; Bräutigam, Christof; Beyer, Anna;
Open Access
English
Published: 23 Jul 2010
Publisher: Zenodo
Abstract
The Webis Query Segmentation Corpus 2010 (Webis-QSeC-10) contains segmentations for 53,437 web queries obtained from Mechanical Turk crowdsourcing (4,850 used for training in our CIKM 2012 paper). For each query, at least 10 MTurk workers were asked to segment the query. The corpus represents the distribution of their decisions. We provide the training and test sets as single folders in Zip archives containing several files. The files "...-queries.txt" contain the query strings and a unique ID for each query. The files "...-segmentations-crowdsourced.txt" contain the crowdsourced segmentations with their number of votes per query ID (see below for an example). The "data" folders contain all the data (n-gram frequencies, PMI values, POS tags, etc.) needed to replicate the evaluation results of our proposed segmentation algorithms. For convenience reasons, the folder "segmentations-of-algorithms" contain the segmentations that our proposed algorithms compute. The original queries were extracted from the AOL query log, and range from 3 to 10 keywords in length. For each query at least 10 MTurk workers were asked to segment the query and their decisions are accumulated in the corpus. The examples below demonstrate two different cases. Sample queries with internal ID (as in "Webis-QSeC-10-training-set-queries.txt"): 2315313155 harvard community credit union 1858084875 women's cycling tops Sample segmentations (as in "webis-qsec-10-training-set-segmentations-crowdsourced.txt"): 2315313155 [(6, 'harvard community credit union'), (2, 'harvard community|credit union'), (1, 'harvard|community|credit union'), (1, 'harvard|community credit union')] 1858084875 [(5, "women's|cycling tops"), (2, "women's|cycling|tops"), (2, "women's cycling|tops"), (1, "women's cycling tops")] Each query has a unique internal ID (e.g., 2315313155 in the first example) and the segmentations file contains at least 10 different decisions the MTurk workers made for that query. In the first example, 6 workers have all 4 keywords in one segment, 2 workers decided to break after the second word (denoted by a |) etc. Note that apostrophe in the second example (query ID 1858084875) is escaped by double quotes around the segmentation strings.
{"references": ["Matthias Hagen, Martin Potthast, Benno Stein, and Christof Br\u00e4utigam. The Power of Na\u00efve Query Segmentation. In Fabio Crestani et al, editors, 33rd International ACM Conference on Research and Development in Information Retrieval (SIGIR 10), pages 797-798, July 2010. ACM. ISBN 978-1-4503-0153-4."]}
Subjects
web queries, query, AOL, segmentation
web queries, query, AOL, segmentation
See an issue? Give us feedback
Related to Research communities
NEANIAS Space Research Community
Download fromView all 2 sources
Do the share buttons not appear? Please make sure, any blocking addon is disabled, and then reload the page.