ICMI 2023 2nd Call for Papers

=================================

ACM ICMI 2023 2ND CALL FOR PAPERS   

=================================

9-13 October 2023, Paris – France

https://icmi.acm.org/2023/

=================================

 

The 25th International Conference on Multimodal Interaction (ICMI 2023) will be held in Paris, France. ICMI is the premier international forum that brings together multimodal artificial intelligence (AI) and social interaction research. Multimodal AI encompasses technical challenges in machine learning and computational modeling such as representations, fusion, data and systems. The study of social interactions englobes both human-human interactions and human-computer interactions.  A unique aspect of ICMI is its multidisciplinary nature which values both scientific discoveries and technical modeling achievements, with an eye towards impactful applications for the good of people and society.

 

This year, the conference is pleased to welcome 3 keynote speakers:

* Maja Mataric (Interaction Lab, University of Southern California)

* Sophie Scott (Institute of Cognitive Neuroscience, University College London)

* Simone Natale (University of Turin)

 

We are keen to showcase novel input and output modalities and interactions to the ICMI community. ICMI 2023 will feature a single-track main conference which includes: keynote speakers, technical full and short papers (including oral and poster presentations), demonstrations, exhibits, doctoral spotlight papers, and late-breaking papers. The conference will also feature workshops and grand challenges. The proceedings of all ICMI 2023 papers, including Long and Short Papers, will be published by ACM as part of their series of International Conference Proceedings and Digital Library, and the adjunct proceedings will feature the workshop papers.

 

Novelty will be assessed along two dimensions: scientific novelty and technical novelty. Accepted papers at ICMI 2023 will need to be novel along one of the two dimensions:

 

* Scientific Novelty: Papers should bring new scientific knowledge about human social interactions, including human-computer interactions. For example, discovering new behavioral markers that are predictive of mental health or how new behavioral patterns relate to children’s interactions during learning. It is the responsibility of the authors to perform a proper literature review and clearly discuss the novelty in the scientific discoveries made in their paper.

 

* Technical Novelty: Papers reviewed with this sub-criterion should include novelty in their computational approach for recognizing, generating or modeling data. Examples include: novelty in the learning and prediction algorithms, in the neural architecture, or in the data representation. Novelty can also be associated with new usages of an existing approach.

 

Please see the Submission Guidelines for Authors https://icmi.acm.org/2023/guidelines-for-authors/ for detailed submission instructions.

 

This year’s conference theme: The theme for this year’s conference is “Science of Multimodal Interactions”. As the community grows, it is important to understand the main scientific pillars involved in deep understanding of multimodal and social interactions. As a first step, we want to acknowledge key discoveries and contributions that the ICMI community enabled over the past 20+ years. As a second step, we reflect on the core principles, foundational methodologies and scientific knowledge involved in studying and modeling multimodal interactions. This will help establish a distinctive research identity for the ICMI community while at the same time embracing its multidisciplinary collaborative nature. This research identity and long-term agenda will enable the community to develop future technologies and applications while maintaining commitment to world-class scientific research.

 

Additional topics of interest include but are not limited to:

 

* Affective computing and interaction

* Cognitive modeling and multimodal interaction

* Gesture, touch and haptics

* Healthcare, assistive technologies

* Human communication dynamics

* Human-robot/agent multimodal interaction

* Human-centered A.I. and ethics

* Interaction with smart environment

* Machine learning for multimodal interaction

* Mobile multimodal systems

* Multimodal behaviour generation

* Multimodal datasets and validation

* Multimodal dialogue modeling

* Multimodal fusion and representation

* Multimodal interactive applications

* Novel multimodal datasets

* Speech behaviours in social interaction

* System components and multimodal platforms

* Visual behaviours in social interaction

* Virtual/augmented reality and multimodal interaction

 

Submissions:

Both comments and pings are currently closed.

Comments are closed.

Design by 2b Consult