~~~~~~~ BRITISH HCI GROUP NEWS SERVICE ~~~~~~~~~~~ ~~ http://www.bcs-hci.org.uk/ ~~ ~~ All news to: [log in to unmask] ~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~ NOTE: Please reply to article's originator, ~~ ~~ not the News Service ~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ WORKSHOP ANNOUNCEMENT AND CALL FOR PAPERS Representing, Annotating, and Evaluating Non-Verbal and Verbal Communicative Acts to Achieve Contextual Embodied Agents http://aos2.uniba.it:8080/aa-ws.html May 29, 2001 Montreal, Canada in conjunction with The Fifth International Conference on Autonomous Agents http://www.csc.liv.ac.uk/~agents2001/ Embodied agents are receiving more and more attention due to the variety of applications, ranging from synthetic actors, pedagogical agents to hearing-impaired helpers. To make the agent interesting to interact with as well as useful for a given application, the agent should behave and talk according to the context where the interaction takes place and to the interlocutor the agent is talking to. Moreover, while talking, verbal and nonverbal signals are at work and should be highly integrated with each others. Interactive systems, then, require a methodology to automatically coordinate verbal and nonverbal signals. Using too simple models of communicative behavior might produce a poor agent with which the User will rapidly feel bored. Several problems arise in attempting to synchronize communication signals with speech: - One needs to define which communicative signals are relevant for a given application to make the agent believable. Some applications require caricature and over-expressive behaviors, while others need very accurate and realistic behaviors. - One needs to consider the context of the interaction. Is the application directed to children, adults, computer scientists? Do we want a friendly agent or an authoritative agent? How to integrate 3D world objects in the context of the conversation (say, objects to point at or to grab while explaining a task, physical objects to play with)? - Having defined these elements, one needs to represent them and find a notation to describe them. In addition, the problem of annotation schemas to be employed in multimodal (verbal and nonverbal) signals has to be solved, that is, how to annotate a discourse also including nonverbal information. MAJOR TOPICS Collecting and using empirical data as a basis for communicative behavior: - how to collect, analyze and exploit empirical data from human-human interaction with regard to verbal / nonverbal inter-dependencies (timing, co-occurrence, context); - how to specify which communicative signals (type of facial expressions, gaze and gestures...) to consider and model for the agent; - how to assess, interpret, and determine responsivenss to the context in which the interaction takes place, that is, which elements of the context are relevant to the conversation and how to represent these elements; - how to design tools for coding, viewing and analyzing multimodal interactions; Representation of multimodal information as a basis for communicative behavior: - how to specify geometric and function properties of real objects for physical embodied agent to interact with; - which models to use for emotions and personality for both the agent and the user (how to represent the ``affect display'' function and/or the potential affective effect of multimodal signals); - how to represent communicative signals and which notations to use (with which type of facial expression, gaze, and gesture as they are manifested); - how to represent the meaning level of communicative signals; - how to represent the knowledge used by the agent; - which discourse annotation should be used to integrate verbal and nonverbal information. - how to ensure synchronization between verbal and nonverbal signals. System architecture and evaluation : - how to define a general architecture to generate agents communicating multimodally; - how to evaluate multimodal interaction with an embodied agent (which modality causes which effect to what degree? Do the modeled communicative functions reach the intended effect?); WORKSHOP FORMAT The workshop will be made of presentations from selected papers. The workshop orientation is more toward computational model and implemented systems rather than purely theoretical statements. Presentation of the theoretical basis of a system could of course be included in the paper but should not be the main and only focus of the paper. The presentations could cover one of the several themes of the workshop and should be oriented toward: - description of the annotating scheme for verbal and non-verbal signals. The presentation should be technically-based rather than purely theoretical. - description of implemented embodied agent systems. The presentation should be on the design and implementation process of these systems. - description of the evaluation scheme used for embodied agent systems. SUBMISSIONS FORMAT Attendance to the workshop will be based on acceptance of paper. Paper submission will be required. Paper length should not extend 6 pages long (using 11pt, single space, all margins of 2cm) and should be accompanied as much as possible with an animation describing the work presented. Every paper submitted will be reviewed by at least 2 reviewers from the program committee. Submissions should be emailed to Catherine Pelachaud ([log in to unmask]) in PS or PDF format by March 16th, 2001. If email is not possible, please sent two copies of your paper to: Catherine Pelachaud University of Rome "La Sapienza" Department of computer science and system via Buonarroti, 12 00185 Rome Italy Note: Workshop participants will be required to register for the Autonomous Agents 2001 main conference. IMPORTANT DATES March 16 - Deadline for paper Submission April 1 - Notification of Acceptance / Rejection April 23 - Deadline for camera-ready paper May 29 - Workshop date ORGANIZING COMMITTEE - Catherine Pelachaud, University of Rome ``La Sapienza'', Italy - Isabella Poggi, University of Rome Tre, Italy PROGRAM COMMITTEE - Norman Badler, University of Pennsylvania, USA - Berardina de Carolis, University of Bari, Italy - James Lester, North Carolina State University, USA - Nadia Magnenat-Thalmann, MiraLab, University of Geneve, Switzerland - Jan-Torsten Milde, University of Bielefeld, Germany - Clifford Nass, Stanford University, USA - Sharon Oviatt, Oregon Graduate Istitute, USA - Catherine Pelachaud, University of Rome ``La Sapienza,'' Italy - Isabella Poggi, University of Rome Tre, Italy - Jeff Rickel, USC Information Sciences Institute, USA - Yasuyuki Sumi, ATR, Japan ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~ To receive HCI news, send the message: ~~ ~~ "JOIN BCS-HCI your_firstname your_lastname" ~~ ~~ to [log in to unmask] ~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~ Newsarchives: ~~ ~~ http://www.jiscmail.ac.uk/lists/bcs-hci.html ~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~ To join the British HCI Group, contact ~~ ~~ [log in to unmask] ~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~