[Asis-l] CFP: Special Issue of Information Retrieval, a Springer journal, on Crowdsourcing for Information Retrieval

Matt Lease ml at ischool.utexas.edu
Sun Nov 14 09:49:01 EST 2010

Information Retrieval, a Springer journal
Special Issue on Crowdsourcing for Information Retrieval

Call for papers: http://ir.ischool.utexas.edu/irj-crowd-cfp.pdf

Submissions due: March 31, 2011

The advent of crowdsourcing is revolutionizing information technology by 
enabling a wide host of new methodology for data collection, annotation, 
and processing, as well as system design and evaluation. Given the 
massive datasets that are both ubiquitous and rapidly growing in today’s 
digital age, the field of information retrieval (IR) stands to 
particularly benefit from such advances as academic researchers and 
industrial practitioners alike leverage the wisdom of crowds to more 
effective cope with and exploit massive datasets. The novelty of 
crowdsourcing makes this is an especially exciting time to focus the IR 
community’s attention toward it and provide an opportunity for 
discussion and education on this key emerging area. Traditionally 
manual-labor intensive processes have been particularly impacted by 
dramatically reducing the time, cost, and effort involved. In IR, this 
has driven a disruptive shift in areas like:

Evaluation: The Cranfield paradigm for evaluating search engines 
requires manually assessing document relevance to search queries. Recent 
work on stochastic evaluation has reduced but not removed this 
dependence on manual assessment.

Supervised Learning: While traditional costs associated with data 
annotation have driven recent machine learning work (e.g. Learning to 
Rank) toward greater use of unsupervised and semi-supervised methods, 
the emergence of crowdsourcing has made labeled data far easier to 
acquire, thereby driving a potential resurgence in supervised or 
semi-supervised methods.

Applications: Crowdsourcing has introduced exciting new opportunities to 
integrate human labor into automated systems: handling difficult cases 
where automation fails, exploiting the breadth of backgrounds, 
geographic dispersion, and real-time crowd response, etc.

While existing studies on crowdsourcing for IR have been encouraging, a 
variety of important questions remain open with regard to theory, 
methodology, policy, and best practices.  See call for additional 
background (http://ir.ischool.utexas.edu/irj-crowd-cfp.pdf).

Call for Papers
The special issue welcomes novel, high-quality manuscripts on the 
development, evaluation, and theoretical analysis of crowdsourcing for 
IR. Submissions to the special issue should not be under consideration 
in any other journal or conference and will be evaluated according to 
the Information Retrieval Journal reviewing criteria and appropriateness 
to the special issue. If the submission is a revision of a previous 
conference paper, the revision must contain significant amplification or 
clarification of the original material or there is some significant 
additional benefit to be gained. For more details, please refer to 
"manuscript submission" on the journal homepage. Submissions should use 
the Information Retrieval Journal style templates available from the 
Journal's homepage and should be submitted through the IR Journal online 
submission page, selecting the “S.I.: Crowdsourcing for Info. Ret.” 
article type.

Topics of interest include but are not limited to the following areas:

* Theoretical, experimental, and/or methodological developments 
advancing state-of-the-art knowledge of crowdsourcing for IR

* Novel applications of IR enabled by crowdsourcing (e.g. use of 
real-time crowd response, using human computation to “close the loop” 
with automation for hybrid IR systems, etc.)

* Tutorials or best practices on employing crowdsourcing to support 
different IR scenarios

* Innovative ideas for better using crowdsourcing platforms such as 
Amazon’s Mechanical Turk, Crowdflower, etc.

* Description of novel software packages improving ease and/or quality 
of crowdsourcing, (e.g. TurkIt, qmturk, Turk Surveyor, etc.)

* Reflective or forward-looking vision on use of crowdsourcing in this area

Important dates
Submissions due: March 31, 2011
Reviewer feedback to authors: May 15, 2011
Revised submissions due: May 31, 2011
Notification of acceptance and final reviewer feedback: June 30, 2011
Final submissions due: July 15, 2011

Guest Editors
All questions regarding submissions should be directed to the special 
issue Guest Editors:

Vitor Carvalho, Microsoft Bing

Matthew Lease, University of Texas at Austin

Emine Yilmaz, Microsoft Research

Matt Lease
Assistant Professor
School of Information and Department of Computer Science
University of Texas at Austin
Voice: (512) 471-9350 · Fax: (512) 471-3971 · Office: UTA 5.450

More information about the Asis-l mailing list