• IEEE.org
  • IEEE CS Standards
  • Career Center
  • About Us
  • Subscribe to Newsletter

0

IEEE
CS Logo
  • MEMBERSHIP
  • CONFERENCES
  • PUBLICATIONS
  • EDUCATION & CAREER
  • VOLUNTEER
  • ABOUT
  • Join Us
CS Logo

0

IEEE Computer Society Logo
Sign up for our newsletter
IEEE COMPUTER SOCIETY
About UsBoard of GovernorsNewslettersPress RoomIEEE Support CenterContact Us
COMPUTING RESOURCES
Career CenterCourses & CertificationsWebinarsPodcastsTech NewsMembership
BUSINESS SOLUTIONS
Corporate PartnershipsConference Sponsorships & ExhibitsAdvertisingRecruitingDigital Library Institutional Subscriptions
DIGITAL LIBRARY
MagazinesJournalsConference ProceedingsVideo LibraryLibrarian Resources
COMMUNITY RESOURCES
GovernanceConference OrganizersAuthorsChaptersCommunities
POLICIES
PrivacyAccessibility StatementIEEE Nondiscrimination PolicyIEEE Ethics ReportingXML Sitemap

Copyright 2025 IEEE - All rights reserved. A public charity, IEEE is the world’s largest technical professional organization dedicated to advancing technology for the benefit of humanity.

FacebookTwitterLinkedInInstagramYoutube
  • Home
  • /Digital Library
  • /Magazine
  • /Is
  • Home
  • / ...
  • /Magazine
  • /Is

CLOSED: Special Issue on Explicable Artificial Intelligence for Affective Computing

IEEE Intelligent Systems seeks submissions for this upcoming special issue.

Important Dates

  • Submission deadline: 1 March 2025
  • Publication date: Sep/Oct 2025


Call for Papers

Background:

As Artificial Intelligence (AI) advances, the need for transparency and interpretability in its decision-making processes becomes more pronounced, especially within the domain of affective computing. The capacity of AI systems to comprehend and react to human emotions introduces ethical considerations, necessitating a delicate equilibrium between innovation and accountability. Various stakeholders, spanning end-users, developers, and policymakers, express a collective need for a more profound comprehension of these systems, particularly in emotionally charged situations.

The motivation of this Special Issue stems from the inherent challenges in creating AI models that not only accurately recognize and respond to human emotions but also provide clear, interpretable insights into their decision-making processes. The Special Issue also aims at enriching the connotation of Explicable AI with diverse and comprehensive dimensions. Expanding the meaning of explicability is not just about deciphering the “black box” nature of AI models; it involves a broader understanding that encapsulates various facets crucial for fostering user trust, ethical considerations, and interdisciplinary collaboration.

Topics:

  • Explainable sentiment analysis, emotion detection, and figurative language processing
  • Neurosymbolic affective computing
  • Multimodal affective computing with explainability
  • Affective intention awareness AI
  • Trustworthy AI for affective computing
  • Affective computing involves multidisciplinary ensemble and explainability
  • Affective computing for science research, e.g., healthcare, education, behavioural, cognitive and social science
  • Granular task decomposition for affective computing
  • Ethical analysis pertains to Explicable AI for affective computing.

Highlights:

The Special Issue will consider papers on the mentioned topics that demonstrate humanitarian value. While achieving state-of-the-art performance is commendable, acceptance priority will be given to works that contribute to the advancement of seven pillars for future AI, including Multidisciplinarity, Task Decomposition, Parallel Analogy, Symbol Grounding, Similarity Measure, Intention Awareness, and Trustworthiness. All submissions to the Special Issue undergo a rigorous editorial pre-screening process to assess their relevance, quality, and originality. This initial screening ensures that the manuscripts align with the thematic focus of the Special Issue and meet the Journal's standards.

Evaluation Criteria:

The evaluation of submitted papers will be guided by the following key questions: 

  1. a) Does the paper contribute to explicable AI in the context of affective computing? 
  2. b) Does the paper provide an adequate level of technical innovation and/or analytical insights?
  3. c) Are the findings or contributions supported by experimental evidence and/or theoretical underpinning? 
  4. d) Is the paper appropriate to be published on IEEE Intelligent Systems?

Peer Review:

The papers will be peer-reviewed by at least three independent reviewers with expertise in the area.


Submission Guidelines

For author information and guidelines on submission criteria, visit the Author’s Information page. Please submit papers through the IEEE Author Portal and be sure to select the special issue or special section name. Manuscripts should not be published or currently submitted for publication elsewhere. Please submit only full papers intended for review, not abstracts, to the ScholarOne portal. If requested, abstracts should be sent by email to the guest editors directly.


Questions? Contact the Lead Guest Editor at rui.mao@ntu.edu.sg

Guest Editors:

Rui Mao (Lead Guest Editor), Nanyang Technological University, Singapore

Erik Cambria, Nanyang Technological University, Singapore

Yang Li, Northwestern Polytechnical University, China

Newton Howard, University of Oxford, UK

LATEST NEWS
MicroLED Photonic Interconnects for AI Servers
MicroLED Photonic Interconnects for AI Servers
Vishkin Receives 2026 IEEE Computer Society Charles Babbage Award
Vishkin Receives 2026 IEEE Computer Society Charles Babbage Award
Empowering Communities Through Digital Literacy: Impact Across Lebanon
Empowering Communities Through Digital Literacy: Impact Across Lebanon
From Isolation to Innovation: Establishing a Computer Training Center to Empower Hinterland Communities
From Isolation to Innovation: Establishing a Computer Training Center to Empower Hinterland Communities
IEEE Uganda Section: Tackling Climate Change and Food Security Through AI and IoT
IEEE Uganda Section: Tackling Climate Change and Food Security Through AI and IoT
Read Next

MicroLED Photonic Interconnects for AI Servers

Vishkin Receives 2026 IEEE Computer Society Charles Babbage Award

Empowering Communities Through Digital Literacy: Impact Across Lebanon

From Isolation to Innovation: Establishing a Computer Training Center to Empower Hinterland Communities

IEEE Uganda Section: Tackling Climate Change and Food Security Through AI and IoT

Blockchain Service Capability Evaluation (IEEE Std 3230.03-2025)

Autonomous Observability: AI Agents That Debug AI

Disaggregating LLM Infrastructure: Solving the Hidden Bottleneck in AI Inference

Get the latest news and technology trends for computing professionals with ComputingEdge
Sign up for our newsletter