----------------We apologize if you receive this
message several times---------
Call for papers:
***** to be held on August 21, 2022 in Montreal *******************************
We are witnessing the
emergence of an “AI economy and society” where AI technologies are increasingly
impacting many aspects of business as well as everyday life. We read with great
interest about recent advances in AI medical diagnostic systems, self-driving
cars, ability of AI technology to automate many aspects of business decisions
like loan approvals, hiring, policing etc. However, AI systems may produce
errors, can exhibit overt or subtle bias, may be sensitive to noise in the
data, and often lack technical and judicial transparency and explainability. These
shortcomings are raising many ethical and policy concerns not only in technical
and academic communities, but also among policymakers and general public, and
will inevitably impede wider adoption of AI in society.
The problems related to
Ethical AI are complex and broad and encompass not only technical issues but
also legal, political and ethical ones. One of the key components of Ethical AI
systems is explainability or transparency, but other issues like detecting
bias, ability to control the outcomes, ability to objectively audit AI systems
for ethics are also critical for successful applications and adoption of AI in
society. Consequently, explainable and Ethical AI are very current and popular
topics in various communities. Our proposed workshop aims to address technical
aspects of explainable and ethical AI in general, and include related
applications and case studies with the aim to address this very important
problems from a broad technical perspective.
The topics comprise but are not limited to:
· Naturally explainable AI
methods
· Post-Hoc Explanation
methods of Deep Neural Networks and Transformers
· Technical issues in AI
ethics including automated audits, detection of bias, ability to control AI
systems to prevent harm and others
· Methods to improve AI
explainability in general, including algorithms and evaluation methods
· User interface and
visualization for achieving more explainable and ethical AI
· Real world applications and
case studies
WorkShop Web Site: https://xaie-icpr.labri.fr
Important Dates:
Paper Submission:
The Proceedings of the EDL-AI 2020 workshop will be
published in the Springer Lecture Notes in Computer Science (LNCS) series.
Papers will be selected by a single blind (reviewers are anonymous) review
process. Submissions must be formatted in accordance with the Springer's
Computer Science Proceedings guidelines . Two types of contribution will be
considered:
Full paper
(12-15 pages)
Short papers
(6-8 pages)
Paper Templates: are
available on the website: https://xaie-icpr.labri.fr
Program Committee:
Alexandre Benoit, France, University Savoie
Mont Blanc / LISTIC
Jenny Benois-Pineau, France, Univ. Bordeaux
/ LaBRI
Romain Bourqui, France, Univ. Bordeaux / LaBRI
André CPLF de Carvalho,
Brazil, University of Sao Paulo / ICMC
Christophe Garcia, France,
LIRIS
Mark Keane, Ireland, UCD
Dublin / Insight SFI Centre for Data Analytics
Harold Mouchere, France,
Université de Nantes / LS2N
Romain Giot, France, Univ. Bordeaux / LaBRI
Théo Jaunet, France, LIRIS
Stefanos Kollias, Greece, National
Technical University of Athens / Image, Video and Multimedia Systems Lab
Noel O’Connor, Ireland, DCU
Dragutin Petkovic, USA, SFSU
Nicolas Thome, France, CNAM/Cedric
Carlos Toxtli, USA, Northeastern University
Jenny
Benois-Pineau, Romain Giot, Romain Bourqui and Dragutin Petkovic
Workshop
Organizers
Unsubscribe:
If you don't already have a password for the LISTSERV.ACM.ORG server, we recommend that you create one now. A LISTSERV password is linked to your email address and can be used to access the web interface and all the lists to which you are subscribed on the LISTSERV.ACM.ORG server.
To create a password, visit:
https://LISTSERV.ACM.ORG/SCRIPTS/WA-ACMLPX.CGI?GETPW1
Once you have created a password, you can log in and view or change your subscription settings at:
https://LISTSERV.ACM.ORG/SCRIPTS/WA-ACMLPX.CGI?SUBED1=MM-INTEREST