ICMI 2017- Proceedings of the 19th ACM International Conference on Multimodal Interaction

Full Citation in the ACM Digital Library

SESSION: Invited Talks

Gastrophysics: using technology to enhance the experience of food and drink (keynote)

Collaborative robots: from action and interaction to collaboration (keynote)

Situated conceptualization: a framework for multimodal interaction (keynote)

Steps towards collaborative multimodal dialogue (sustained contribution award)

SESSION: Oral Session 1: Children and Interaction

Tablets, tabletops, and smartphones: cross-platform comparisons of children’s touchscreen interactions

Toward an efficient body expression recognition based on the synthesis of a neutral movement

Interactive narration with a child: impact of prosody and facial expressions

Comparing human and machine recognition of children’s touchscreen stroke gestures

SESSION: Oral Session 2: Understanding Human Behaviour

Virtual debate coach design: assessing multimodal argumentation performance

Predicting the distribution of emotion perception: capturing inter-rater variability

Automatically predicting human knowledgeability through non-verbal cues

Pooling acoustic and lexical features for the prediction of valence

SESSION: Oral Session 3: Touch and Gesture

Hand-to-hand: an intermanual illusion of movement

An investigation of dynamic crossmodal instantiation in TUIs

“Stop over there”: natural gesture and speech interaction for non-critical spontaneous intervention in autonomous driving

Pre-touch proxemics: moving the design space of touch targets from still graphics towards proxemic behaviors

Freehand grasping in mixed reality: analysing variation during transition phase of interaction

Rhythmic micro-gestures: discreet interaction on-the-go

SESSION: Oral Session 4: Sound and Interaction

Evaluation of psychoacoustic sound parameters for sonification

Utilising natural cross-modal mappings for visual control of feature-based sound synthesis

SESSION: Oral Session 5: Methodology

Automatic classification of auto-correction errors in predictive text entry based on EEG and context information

Cumulative attributes for pain intensity estimation

Towards the use of social interaction conventions as prior for gaze model adaptation

Multimodal sentiment analysis with word-level fusion and reinforcement learning

IntelliPrompter: speech-based dynamic note display interface for oral presentations

SESSION: Oral Session 6: Artificial Agents and Wearable Sensors

Head and shoulders: automatic error detection in human-robot interaction

The reliability of non-verbal cues for situated reference resolution and their interplay with language: implications for human robot interaction

Do you speak to a human or a virtual agent? automatic analysis of user’s social cues during mediated communication

Estimating verbal expressions of task and social cohesion in meetings by quantifying paralinguistic mimicry

Data augmentation of wearable sensor data for parkinson’s disease monitoring using convolutional neural networks

SESSION: Poster Session 1

Automatic assessment of communication skill in non-conventional interview settings: a comparative study

Low-intrusive recognition of expressive movement qualities

Digitising a medical clerking system with multimodal interaction support

GazeTap: towards hands-free interaction in the operating room

Boxer: a multimodal collision technique for virtual objects

Trust triggers for multimodal command and control interfaces

TouchScope: a hybrid multitouch oscilloscope interface

A multimodal system to characterise melancholia: cascaded bag of words approach

Crowdsourcing ratings of caller engagement in thin-slice videos of human-machine dialog: benefits and pitfalls

Modelling fusion of modalities in multimodal interactive systems with MMMM

Temporal alignment using the incremental unit framework

Multimodal gender detection

How may I help you? behavior and impressions in hospitality service encounters

Tracking liking state in brain activity while watching multiple movies

Does serial memory of locations benefit from spatially congruent audiovisual stimuli? investigating the effect of adding spatial sound to visuospatial sequences

ZSGL: zero shot gestural learning

Markov reward models for analyzing group interaction

Analyzing first impressions of warmth and competence from observable nonverbal cues in expert-novice interactions

The NoXi database: multimodal recordings of mediated novice-expert interactions

Head-mounted displays as opera glasses: using mixed-reality to deliver an egalitarian user experience during live events

SESSION: Poster Session 2

Analyzing gaze behavior during turn-taking for estimating empathy skill level

Text based user comments as a signal for automatic language identification of online videos

Gender and emotion recognition with implicit user signals

Animating the adelino robot with ERIK: the expressive robotics inverse kinematics

Automatic detection of pain from spontaneous facial expressions

Evaluating content-centric vs. user-centric ad affect recognition

A domain adaptation approach to improve speaker turn embedding using face representation

Computer vision based fall detection by a convolutional neural network

Predicting meeting extracts in group discussions using multimodal convolutional neural networks

The relationship between task-induced stress, vocal changes, and physiological state during a dyadic team task

Meyendtris: a hands-free, multimodal tetris clone using eye tracking and passive BCI for intuitive neuroadaptive gaming

AMHUSE: a multimodal dataset for HUmour SEnsing

GazeTouchPIN: protecting sensitive data on mobile devices using secure multimodal authentication

Multi-task learning of social psychology assessments and nonverbal features for automatic leadership identification

Multimodal analysis of vocal collaborative search: a public corpus and results

UE-HRI: a new dataset for the study of user engagement in spontaneous human-robot interactions

Mining a multimodal corpus of doctor’s training for virtual patient’s feedbacks

Multimodal affect recognition in an interactive gaming environment using eye tracking and speech signals

SESSION: Demonstrations 1

Multimodal interaction in classrooms: implementation of tangibles in integrated music and math lessons

Web-based interactive media authoring system with multimodal interaction

Textured surfaces for ultrasound haptic displays

Rapid development of multimodal interactive systems: a demonstration of platform for situated intelligence

MIRIAM: a multimodal chat-based interface for autonomous systems

SAM: the school attachment monitor

The Boston Massacre history experience

Demonstrating TouchScope: a hybrid multitouch oscilloscope interface

The MULTISIMO multimodal corpus of collaborative interactions

Using mobile virtual reality to empower people with hidden disabilities to overcome their barriers

SESSION: Demonstrations 2

Bot or not: exploring the fine line between cyber and human identity

Modulating the non-verbal social signals of a humanoid robot

Thermal in-car interaction for navigation

AQUBE: an interactive music reproduction system for aquariums

Real-time mixed-reality telepresence via 3D reconstruction with HoloLens and commodity depth sensors

Evaluating robot facial expressions

Bimodal feedback for in-car mid-air gesture interaction

A modular, multimodal open-source virtual interviewer dialog agent

Wearable interactive display for the local positioning system (LPS)

SESSION: Grand Challenge

From individual to group-level emotion recognition: EmotiW 5.0

Multi-modal emotion recognition using semi-supervised learning and multiple neural networks in the wild

Modeling multimodal cues in a deep learning-based framework for emotion recognition in the wild

Group-level emotion recognition using transfer learning from face identification

Group emotion recognition with individual facial emotion CNNs and global image based CNNs

Learning supervised scoring ensemble for emotion recognition in the wild

Group emotion recognition in the wild by combining deep neural networks for facial expression classification and scene-context analysis

Temporal multimodal fusion for video emotion classification in the wild

Audio-visual emotion recognition using deep transfer learning and multiple temporal models

Multi-level feature fusion for group-level emotion recognition

A new deep-learning framework for group emotion recognition

Emotion recognition in the wild using deep neural networks and Bayesian classifiers

Emotion recognition with multimodal features and temporal models

Group-level emotion recognition using deep models on image scene, faces, and skeletons

SESSION: Doctoral Consortium

Towards designing speech technology based assistive interfaces for children's speech therapy

Social robots for motivation and engagement in therapy

Immersive virtual eating and conditioned food responses

Towards edible interfaces: designing interactions with food

Towards a computational model for first impressions generation

A decentralised multimodal integration of social signals: a bio-inspired approach

Human-centered recognition of children's touchscreen gestures

Cross-modality interaction between EEG signals and facial expression

Hybrid models for opinion analysis in speech interactions

Evaluating engagement in digital narratives from facial data

Social signal extraction from egocentric photo-streams

Multimodal language grounding for improved human-robot collaboration: exploring spatial semantic representations in the shared space of attention

SESSION: Workshop Summaries

ISIAA 2017: 1st international workshop on investigating social interactions with artificial agents (workshop summary)

WOCCI 2017: 6th international workshop on child computer interaction (workshop summary)

MIE 2017: 1st international workshop on multimodal interaction for education (workshop summary)

Playlab: telling stories with technology (workshop summary)

MHFI 2017: 2nd international workshop on multisensorial approaches to human-food interaction (workshop summary)