~~~~~~~ BRITISH HCI GROUP NEWS SERVICE ~~~~~~~~~~~
~~ http://www.bcs-hci.org.uk/ ~~
~~ All news to: [log in to unmask] ~~
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
~~ NOTE: Please reply to article's originator, ~~
~~ not the News Service ~~
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
3-Day Workshop on
EVALUATING COLLABORATIVE ENTERPRISES
at the IEEE 11th International Workshop on Enabling Technologies:
Infrastructure for Collaborative Enterprises Carnegie Mellon
University, Pittsburgh, PA, USA June 10-12, 2002
WETICE is an annual, international forum for state-of-the-art research
in enabling technologies for collaboration, consisting of a number of
workshops on different topics related to collaboration technology.
WETICE workshops are not an "add-on" - they are the main activity of the
conference, and are designed to promote fruitful discussions on the
latest technology developments, directions, problems, and requirements.
WETICE main web page: http://netserver.cerc.wvu.edu/wetice02
3-Day Workshop on EVALUATING COLLABORATIVE ENTERPRISES
======================================================
Workshop Co-Chairs:
Elaine M. Raybourn
Sandia National Laboratories, USA
Julian Newman
Glasgow Caledonian University, UK
Effective collaboration involves people, communication, and the
co-creation of meaning through information sharing that may be
synchronous and/or asynchronous. Internet technology has
enabled basic communication infrastructures to facilitate people working
together collaboratively over the Web. However, using these applications
has
proven much more difficult in enabling effective collaboration within
and among enterprises than previously acknowledged. Researchers and
practioners need tools to measure the incremental progress towards
developing useful collaborative groupware systems, as well
as methods to evaluate the impact of specific technologies on the
effectiveness of human to human, or human to machine collaboration. We
believe developing effective evaluation methodologies will facilitate
progress in designing and deploying Web-based collaborative
technologies.
The primary goal of this workshop is to provide a forum, in which
researchers and practitioners can share tools and evaluation
methodologies of collaborative enterprises, lessons learned from
deployment of collaborative technologies in organizations or
educational institutions, and ideas for directions the area of
evaluation must move toward in order to facilitate the progress of
distributed collaboration.
This workshop is an excellent opportunity to bring together
people who are addressing the unique and challenging needs of
collaborative
enterprise evaluation. We welcome delegates from all aspects of
industry and academia. Previous WETICE workshops have attracted
delegates from areas such as computer science, information science,
artificial intelligence, communication, psychology, sociology,
education, human factors, usability, systems engineering,
and library sciences.
Important Dates
Full papers due to individual WETICE workshops - March 29, 2002
Notification to authors - April 26, 2002
Final papers for WETICE Post-proceedings - May 22
2002 Advance registration deadline - May 22, 2002
Final workshop reports by Workshop Chairs - June 28, 2002 (after the
workshop)
Workshop Dates - June 10-12, 2002
Workshop Description:
The Evaluating Collaborative Enterprises workshop builds upon the
research presentations and the vibrant discussion from last year's
debut of this workshop. This year we intend to kick off
discussions by providing an online discussion forum before the workshop.
A common theme identified at WETICE 2001 was the need for research
and practice to better understand the characteristics of the
collaborative
enterprise of interest, and determine what evaluation approaches,
methods, techniques, tools, and associated metrics are most
appropriate. Ultimately, there needs to be a framework or taxonomy
that can answer:
+ Which approaches are best for evaluating different types of
collaborative systems? Can these be categorized?
+ In the design cycle of collaborative software
development, when are particular evaluation approaches effective and
when
are they not? Can a spectrum be developed?
+ What combination of methods and techniques for gathering metrics are
most effective for the situation under evaluation?
+ Which evaluation approaches, methods and techniques address
collaboration process and product effectiveness, efficiency and
satisfaction?
+ Which metrics address product and process effectiveness,
efficiency, and satisfaction of the collaborative enterprise under
study?
+ What evaluation tools and mechanisms are best for generating specific
metrics?
Topics that contribute to this framework may include:
+ Benchmark collaboration scenarios and associated evaluation measures
for groupware system design and development,
+ Adaptation of single-user software development and evaluation
techniques
to groupware evaluation,
+ Groupware design principles or heuristics for use in groupware
evaluation.
+ Analysis of group characteristics (organizational, behavioral, and
technical) and corresponding groupware characteristics,
+ Collaboration evaluation methods and tools that use design
ethnography,
+ Case studies evaluating collaborative enterprises,
+ Evaluating cross-cultural techniques and solutions,
+ Methods and tools for lowering the cost of evaluating collaborative
enterprises,
+ Methods and tools for effective field study evaluation.
Submission of Technical papers:
Authors of technical papers should submit an original paper (not
submitted or published elsewhere) in portable document format (pdf)
format by email to [log in to unmask] <mailto:[log in to unmask]>.
Submissions should include the title of the paper, the name and
affiliation of each author, a 150-word abstract, and no more than
eight keywords. Submissions should not exceed 3000 words in length
(including figure equivalents). The name, position, address, telephone
number, fax number and e-mail address of the author responsible for
correspondence must be included.
Submission of Poster papers or Demonstrations:
The Evaluating Collaborative Enterprises workshop committee invites
poster papers or demonstrations of new and exciting research work,
work-in-progress and demonstrations related to the workshop themes.
Authors of poster papers should submit an extended abstract
by email to [log in to unmask] <mailto:[log in to unmask]>.
Submissions
should include the title of the paper the name and
affiliation of each author, a 150-word abstract, and no more than four
keywords.
Submissions should not exceed 1500 words in length. The name,
position, address, telephone number, fax number and e-mail
address of the author responsible for correspondence must be included.
A selection of accepted papers will be published by the IEEE Computer
Society in the workshop post-proceedings. Papers accepted for
publication in the proceedings are limited to six pages (approximately
2000-2500 words) in IEEE format (two columns, single-spaced, 10pt Times)
for technical papers
and two pages for poster papers. Authors are strongly encouraged to
adhere to this format also when submitting papers to the workshop.
Detailed information on the IEEE format (together with some
templates) is available at http://www.computer.org/cspress/instruct.htm.
Submission of Panel proposals:
Send six copies of panel proposals, or by email, to
[log in to unmask] <mailto:[log in to unmask]>. Include a title, a
150-word scope statement, proposed session chair, panelists and their
affiliations,
the organizer's affiliation, address, telephone number, fax number,
and e-mail address.
Workshop Program Committee
Janet Allen, Georgia Tech University, USA
Robert Allen, NIST, USA
Jeff Campbell, University of Maryland, Baltimore County, USA
Elizabeth Churchill, FX Palo Alto Laboratory, USA
Jill Drury, MITRE Corporation, USA
Grainne Conole, University of Bristol, UK
Nick Kings, BTexact, UK
David McDonald, University of Washington, USA
Emile Morse, NIST, USA
Alan Munro, Strathclyde University, UK
Hilary Palmen, Microsoft Corporation, USA
Wolfgang Prinz, Fraunhofer FIT-CSCW, Germany
Tim Regan, Microsoft Corporation, USA
Mickey Potts Steves, NIST, USA
Christine Yang, Sandia National Laboratories, USA
Please send all inquiries regarding this workshop to the Workshop
Co-chairs. For inquiries regarding WET ICE in general, contact
[log in to unmask] or call (U.S.) +1-304-293-7226.
Workshop Program Co-Chairs
Elaine M. Raybourn
Computational Initiatives
Sandia National Laboratories
P.O. Box 5800, MS 1188
Albuquerque, New Mexico 87185 USA
E-mail: [log in to unmask]
Julian Newman
Department of Computing
Glasgow Caledonian University
Cowcaddens Road
Glasgow G4 0BA
Scotland
UK
E-mail: [log in to unmask]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
~~ To receive HCI news, send the message: ~~
~~ "JOIN BCS-HCI your_firstname your_lastname" ~~
~~ to [log in to unmask] ~~
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
~~ Newsarchives: ~~
~~ http://www.jiscmail.ac.uk/lists/bcs-hci.html ~~
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
~~ To join the British HCI Group, contact ~~
~~ [log in to unmask] ~~
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
________________________________________________________________________________
This email (and any attachment) is private and confidential. If you were
not meant to receive it, please delete it and tell us by phone or email.
Although we scan all email and attachments we do not accept responsibility
for viruses.
________________________________________________________________________________
|