CHI-ANNOUNCEMENTS Archives

ACM SIGCHI General Interest Announcements (Mailing List)

CHI-ANNOUNCEMENTS@LISTSERV.ACM.ORG

Options: Use Classic View

Use Proportional Font
Show Text Part by Default
Show All Mail Headers

Topic: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
"Marvel, Jeremy A. (Fed)" <[log in to unmask]>
Thu, 25 Feb 2021 20:18:47 +0000
text/plain (75 lines)
Apologies for cross-posting.

Due to a number of requests, we have extended the submission deadline for the extended abstracts.  Submissions will be accepted through March 5, 2021, and notifications of acceptance will be made on a rolling basis.

The workshop schedule has been updated on the website.

Workshop website:  https://hri-methods-metrics.github.io/
Topics sought:

  *   Abstracts of systems, test methods, data sets, replicability studies, and metrics for evaluating HRI in human-robot teams
  *   Abstracts of proposals for repeatability/replicability studies in HRI
Important dates:

  *   5 March 2021: Submission Deadline for Extended Abstracts
  *   12 March 2021: Full-day Workshop

Workshop Description:

Despite large advances in robot interfaces and user-centric robot designs, the need for effective HRI continues to present challenges for the field of robotics. A key barrier to achieving effective human-robot teaming in a multitude of domains is that there are few consistent test methods and metrics for assessing HRI effectiveness. The necessity for validated metrology is driven by the desire for repeatable and consistent evaluations of HRI methodologies.

This full-day workshop at the 2021 ACM/IEEE HRI Conference<https://humanrobotinteraction.org/2021/> will address the issues surrounding the development of novel test methods and metrics for evaluating HRI performance across the multitude of system and application domains, including industrial, social, medical, field and service robotics. This workshop is driven by the need for establishing consistent standards for evaluating HRI in real-world applications, and how new HRI technologies are being tested and evaluated to ensure performance. Specific goals include the following:


  *   to develop and encourage the use of consistent metrology for HRI, producing quality data sets of pragmatic applications, and validating human subject studies for HRI;
  *   to explore novel and emerging \metrology tools that have broad applicability across HRI domains;
  *   to encourage the creation and sharing of high-quality, consistently formatted datasets for HRI research; and
  *   to promote the development of reproducible, metrics oriented studies that seek to understand and model the human element of HRI teams.

Presentations by contributing authors will focus on the documentation of the test methods, metrics, and data sets used in their respective studies. Keynote and invited speakers will be selected from a targeted list of HRI researchers across a broad spectrum of application domains. Poster session participants will be selected from contributors reporting late-breaking evaluations and their preliminary results.

Discussions are intended to highlight the various approaches, requirements, and opportunities of the research community toward assessing HRI performance, enabling advances in HRI research, and establishing trust in HRI technologies. Specific topics of discussion will include:

  *   reproducible and repeatable studies with quantifiable test methods and metrics;
  *   human-robot collaboration and teaming test methods;
  *   human data set content transferability, and traceability;
  *   HRI metrics (e.g., situation and cultural awareness);
  *   industry-specific metrology requirements.

A workshop report documenting the presentations, discussions, and ensuing take-away and action items will be produced, and made publicly available.

Finally, this workshop is the third in a series of workshops leading toward formalized HRI performance standards. The IEEE Robotics and Automation Society (RAS) is hosting and supporting this standardization effort<https://www.ieee-ras.org/industry-government/standards/active-projects/metrology-for-human-robot-interaction>. Early workshops are intended to target community and consensus building, and on the establishment of a culture of repeatable and reproducible, metrology-based research in HRI. A fourth workshop is planned for the 2022 ACM/IEEE International Conference on Human Robot Interaction, and will specifically address the action items identified in this year's workshop.

With best regards,
Dr. Jeremy A. Marvel, National Institute of Standards and Technology (NIST), USA
Shelly Bagchi, National Institute of Standards and Technology (NIST), USA
Megan Zimmerman, National Institute of Standards and Technology (NIST), USA
Murat Aksu, National Institute of Standards and Technology (NIST), USA
Brian Antonishek, National Institute of Standards and Technology (NIST), USA
Dr. Yue Wang, Clemson University
Dr. Ross Mead, Semio
Dr. Terry Fong, National Aeronautics and Space Administration (NASA), USA
Dr. Heni Ben Amor, Arizona State University


Jeremy A. Marvel, Ph.D.
Computer Scientist, Project Leader Performance of Human-Robot Interaction
U.S. Department of Commerce
National Institute of Standards and Technology
Engineering Laboratory, Intelligent Systems Division
100 Bureau Drive, Stop 8230
Gaithersburg, MD 20899 USA
Tel:  301-975-4592
Fax:  301-990-9688
[log in to unmask]<mailto:[log in to unmask]>



    ----------------------------------------------------------------------------------------
    To unsubscribe from CHI-ANNOUNCEMENTS send an email to:
     mailto:[log in to unmask]

    To manage your SIGCHI Mailing lists or read our polices see:
     https://sigchi.org/operations/listserv/
    ----------------------------------------------------------------------------------------

ATOM RSS1 RSS2