Connectionists: [CFP] Special Session IJCNN 2023 - The Coming of Age of Explainable AI (XAI) and ML

Alfredo Vellido avellido at cs.upc.edu
Wed Jan 18 03:50:42 EST 2023


Apologies for cross-posting
===================
1st CALL FOR PAPERS
===================

IEEE IJCNN 2023 Special Session on
The Coming of Age of Explainable AI (XAI) and Machine Learning

June 18-23, 2023, Queensland, Australia
www.cs.upc.edu/~avellido/research/conferences/IJCNN2023-XAIcomingofage.html
https://2023.ijcnn.org/authors/paper-submission

Aims & Scope
------------
Much of current research on Machine Learning (ML) is dominated by 
methods of the Deep Learning family. The more complex their 
architectures, the more difficult the interpretation or explanation of 
how and why a particular network prediction is obtained, or the 
elucidation of which components of the complex system contributed 
essentially to the obtained decision. This brings about the concern 
about interpretability and non-transparency of complex models, 
especially in high-stakes applications areas such as healthcare, 
national security, industry or public governance, to name a few, in 
which decision making processes may affect citizens. This is, for 
instance, made especially relevant by rapid developments in the field of 
autonomous systems – from cars that drive themselves to partner robots 
and robotic drones. DARPA (Defense Advanced Research Projects Agency), a 
research agency of US Department of Defense, was the first to start a 
research program on Explainable AI 
(https://www.darpa.mil/program/explainable-artificial-intelligence) with 
the goal “to create a suite of machine learning techniques that (1) 
Produce more explainable models, while maintaining a high level of 
learning performance (prediction accuracy); and (2) Enable human users 
to understand, appropriately trust, and effectively manage the emerging 
generation of artificially intelligent partners.” Research on 
Explainable AI (XAI) is now supported worldwide by a variety of public 
institutions and legal regulations, such as European Union’s General 
Data Protection Regulation (GDPR) and the forthcoming Artificial 
Intelligence Act. Similar concerns about transparency and 
interpretability are being raised by governments and organizations 
worldwide. The lack of transparency (interpretability and 
explainability) of many ML approaches in the light of regulations may 
end up limiting ML to niche applications and poses a significant risk of 
costly mistakes without the mitigation of a sound
understanding about the flow of information in the model.
For this special session, we invite papers that address many of the 
challenges of XAI in the context of ML models and algorithms. We are 
interested in papers on efficient and innovative algorithmic approaches 
to XAI and their actual applications all areas. This session also aims 
to explore the performance-versus-explanation trade-off space for 
high-stakes applications of ML in light of all types of AI regulation. 
Comprehensive survey papers on existing technologies for XAI are also 
welcome. We aim to bring together researchers from different fields to 
discuss key issues related to the research and applications of XAI 
methods and to share their experiences of solving problems in 
high-stakes applications in all domains.
Topics that are of interest to this session include but are not limited to:

  New neural network architectures and algorithms for XAI
  Interpretability by design
  Rule extraction algorithms for deep neural networks
  Augmentations of AI methods to increase interpretability and transparency
  Innovative applications of XAI
  Verification of AI performance
  Regulation-compliant XAI methods
  Explanation-generation methods for high-stakes applications
  Stakeholder-specific XAI methods for high-stakes applications
  XAI methods auditing in specific domains
  Human-in-the-loop ML: bridging the gap between data scientists and 
end-users
  XAI through Data Visualization
  Interpretable ML pipelines
  Query Interfaces for DL
  Active and Transfer learning with transparency
  Relevance and Metric Learning
  Deep Neural Reasoning
  Interfaces with Rule-Based Reasoning, Fuzzy Logic and Natural Language 
Processing

Important Dates
---------------
Paper submission: January 31, 2023 (likely to be extended)
Paper decision notification: March 31, 2023

Session Chairs
--------------
Qi Chen, Victoria University of Wellington, New Zealand
José M Juárez, Universidad de Murcia, Spain
Paulo Lisboa, Liverpool John Moores University, U.K.
Asim Roy, Arizona State University, U.S.A.
Alfredo Vellido, Universitat Politècnica de Catalunya, Spain



More information about the Connectionists mailing list