ACM SIGCHI General Interest Announcements (Mailing List)


Options: Use Forum View

Use Monospaced Font
Show Text Part by Default
Condense Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
text/plain; charset=windows-1252
Tue, 15 May 2012 10:16:14 +0200
Pablo Castells <[log in to unmask]>
"ACM SIGCHI General Interest Announcements (Mailing List)" <[log in to unmask]>
Pablo Castells <[log in to unmask]>
text/plain (154 lines)
------------------ 2nd Call for Papers - RUE 2012 ------------------
                International ACM RecSys Workshop on
      Recommendation Utility Evaluation: Beyond RMSE - RUE 2012
                 Dublin, Ireland, 9 September 2012

* Submission deadline: 8 June 2012 *


Measuring the error in rating value prediction has been by far the dominant
evaluation methodology in the Recommender Systems literature. Yet there
seems to be a general consensus that this criterion alone is far from being
enough to assess the practical effectiveness of a recommender system in
matching user needs. The end users of recommendations receive lists of
items rather than rating values, whereby recommendation accuracy metrics
-as surrogates of the evaluated task- should target the quality of the item
selection, rather than the numeric system scores that determine this
selection. Gaps in the adoption of ranking evaluation methodologies (e.g.
IR metrics) result in methodological divergences though, which hinder the
interpretation and comparability of empirical observations by different

On the other hand, accuracy is only one among several relevant dimensions
of recommendation effectiveness. Novelty and diversity, for instance, have
been recognized as key aspects of recommendation utility in many
application domains. From the business point of view, the value added by
recommendation can be measured more directly in terms of clickthrough,
conversion rate, order size, returning customers, increased revenue, etc.
Furthermore, web portals and social networks commonly face multiple
objective optimization problems related to user engagement, requiring
appropriate evaluation methodologies for optimizing along the entire
recommendation funnel. Other potentially relevant dimensions of effective
recommendations for consumers and providers include confidence, coverage,
risk, cost, robustness, usability, etc.

While the need for further extension, formalization, clarification and
standardization of evaluation methodologies is recognized in the community,
this need is still unmet for a large extent. When engaging in evaluation
work, researchers and practitioners are still often faced with experimental
design questions for which there are currently not always precise and
consensual answers. RUE 2012 aims to gather researchers and practitioners
interested in developing better, clearer, and/or more complete evaluation
methodologies for recommender systems -or just seeking clear guidelines for
their experimental needs. The workshop aims to provide an informal setting
for exchanging and discussing ideas, sharing experiences and viewpoints,
seeking to advance in the consolidation and convergence of experimental
methods and practice.


We invite the submission of papers reporting original research, studies,
advances, experiences, or work in progress in the scope of recommender
system utility evaluation. The topics the workshop seeks to address include
–though need not be limited to– the following:

* Recommendation quality dimensions
  - Effective accuracy, ranking quality
  - Novelty, diversity, unexpectedness, serendipity
  - Utility, gain, cost, risk, benefit
  - Robustness, confidence, coverage, ease of use, etc.
* Matching metrics to tasks, needs, and goals
  - User satisfaction, user perception, human factors
  - Business-oriented evaluation
  - Multiple objective optimization, user engagement
  - Quality of service, quality of experience
* Evaluation methodology and experimental design
  - Definition and evaluation of new metrics, studies of existing ones
  - Adaptation of methodologies from related fields: IR, Machine Learning,
HCI, etc.
  - Evaluation theory
* Practical aspects of evaluation
  - Offline and online experimental approaches
  - Simulation-based evaluation
  - Datasets and benchmarks
  - Validation of metrics


Two submission types are accepted: technical papers of up to 6 pages, and
position papers up to 3 pages. Each paper will be evaluated by at least two
reviewers from the Programme Committee. The papers will be evaluated for
their originality, contribution significance, soundness, clarity, and
overall quality. Within a required quality standard, position papers will
be appreciated for presenting new perspectives and insights, and their
potential for provoking thought and stimulating discussion.

All submissions shall adhere to the standard ACM SIG proceedings format: The accepted
papers will be published in the CEUR Proceedings series.

Submissions shall be sent as a pdf file through the online submission
system is now open at:

Important dates

Paper submission deadline: 8 June 2012
Author notification:       22 June 2012
Camera ready version due:  6 July 2012
RUE 2012 workshop:         9 September 2012

Programme Committee

Gediminas Adomavicius, University of Minnesota, USA
Alejandro Bellogín, Universidad Autónoma de Madrid, Spain
Iván Cantador, Universidad Autónoma de Madrid, Spain
Licia Capra, University College London, UK
Ňscar Celma, Gracenote, USA
Charles Clarke, University of Waterloo, Canada
Paolo Cremonesi, Politecnico di Milano, Italy
Juan Manuel Fernández-Luna, Universidad de Granada, Spain
Pankaj Gupta, Twitter, USA
Juan F. Huete, Universidad de Granada, Spain
Dietmar Jannach, University of Dortmund, Germany
Jaap Kamps, University of Amsterdam, Netherlands
Neal Lathia, University College London, UK
Jérôme Picault, Bell Labs, Alcatel-Lucent, France
Filip Radlinski, Microsoft, Canada
Francesco Ricci, Free University of Bozen-Bolzano, Italy
Fabrizio Silvestri, Consiglio Nazionale delle Ricerche, Italy
Harald Steck, Netflix, USA
David Vallet, Universidad Autónoma de Madrid, Spain
Paulo Villegas, Telefónica R&D, Spain
Jun Wang, University College London, UK
Yi Zhang, University of California, Santa Cruz, USA


Xavier Amatriain, Netflix, USA
Pablo Castells, Universidad Autónoma de Madrid, Spain
Arjen de Vries, Centrum Wiskunde & Informatica, Netherlands
Christian Posse, LinkedIn, USA

Contact email: [log in to unmask]

More info at:

    For news of CHI books, courses & software, join CHI-RESOURCES
     mailto: [log in to unmask]

    To unsubscribe from CHI-ANNOUNCEMENTS send an email to
     mailto:[log in to unmask]

    For further details of CHI lists see