Friends,
Ranulph, Chris, and Catherine each state important points. I agree
with them - conferences serve more purposes than simply publishing or
exposing one's own research. For my school, presenting at a
conference is an effective mechanism for determining who cares enough
about attending to do the work of submitting a presentation. It
enables us to fly the flag and therefore justifies the cost. It is
part of the publishing mechanism but not central. Nevertheless,
conferences are vital for other reasons.
To cut conference funding and conference attendance would diminish
the general development of research networks and the entire research
agenda.
That said, I've fairly well shifted my attention from large
conferences to small, focused events where it is possible to "confer"
in dialogue with other participants rather than rush from room to
room finding out what the program contains. This past year I
co-chaired Wonderground, but I did not do it because I like big
conferences. I did it because I like working with Eduardo, Martim,
and Terry. My preference is small, focused conferences such as the
forthcoming conference on events and event structures at Denmark's
Design School.
Conferences -- even large ones -- nevertheless serve many purposes.
They give people a chance to meet and to survey the field. They
function as a kind of ecological wetlands between research at home
and formal publication, and they give people a chance to gather
responses and gain insight on one's own work and on the field. In
some places, conferences also serve as a training ground for doctoral
candidates and inexperienced or younger researchers to begin
presenting their work. These deserve discussion. Eliminating
conference support is yet another unfortunate outcome of the
audit-driven mentality that finds the reasons and purposes of
conferences to diffuse to measure in quantitative terms.
Can conferences be more fruitful to serve the field better -- our
field, any field? Yes. That is a specific thread in its own right.
Despite the problems that pop up in specific conferences or the uses
to which some put them, the medium of the conference remains relevant
to design research as it does to most fields.
Terry's note on the use of specific journals as a metric deserves a
further comment. This kind of policy will prove to be a real problem
on a national level. In some fields, some schools establish lists of
target journals that serve for tenure and promotion. Only these
journals count. This is especially the case in some North American
universities. On the one hand, publishing guarantees visibility, but
the fact that citation rankings occur to and from journals in only
one data base -- without including citations to books or monographs
-- means that impact ratings do not measure true impact on a field,
only impact between and among journals in the database.
This gets sillier still when one realizes that journals have a
limited number of pages per year. This places a limit on the number
of possible articles that any journal can publish. In a field of
10,000 scholars with only one journal, only a couple dozen scholars
-- and only a couple dozen schools -- will have any "impact" in a
given year. A single university can afford -- albeit mistakenly -- to
pursue the "top 5" journals strategy. An entire nation cannot, at
least not if it wishes to bring all its universities and research
centers up, rather than to use a proxy measure that will inevitably
sort out all but a handful of research universities, relegating the
rest to second-class or third-class ranking. It seems to me that
Australia's policy will have the effect of sorting out status among
Australian universities, reinforcing the dominant status of the "big
eight" while reducing the standing of the other 30 or so Australian
universities. In international terms, this will be a disaster,
especially since education is one of Australia's most profitable
international industries.
One can use journals in a metric system, but it must be reasonable
and appropriately wide to encourage research and raise the research
capacity of the nation as a whole. This means using a metric that
genuinely covers the fields it measures, and allows opportunity to
all universities rather than an elite few. The elite will remain
elite. The others deserve a chance to improve.
In Norway, the government uses a publishing metric to measure
research productivity. The government maintains a database of
research publishing channels. The database includes journals and
academic or scientific publishers, with all approved channels on
level 1 and some select channels on level 2. A journal article is
worth 1 point or 3 points respectively, a monograph is worth 5 points
or 8 points, and a book chapter is worth 7/10 point or 1 point. Each
publication by a faculty member accumulates a number of points and
each school tallies up and presents all its publishing points in each
publishing year. The ministry allocates basic research funds to
research-based universities, university colleges, and professional
schools based on the total number of points.
Our database shows 85 journals with the word "design" in the title.
We admittedly miss some design and design research journals that do
not use the word - the new journal Artifact is an example. (Artifact
is registered, but it does not show up in searching design as a title
word.) Nevertheless, the broad spectrum covers much of the field. If
you would like to see for yourself, go to URL:
http://dbh.nsd.uib.no/kanaler/
No system is perfect, but a broad system, works well enough. It
counts respectable journals while placing greater emphasis on a
smaller selection of elite journals.
For those who wish to look further into citations and impact, I would
suggest visiting the section of William Starbuck's web site titled
"Citations of Journals Related to Business." Go to URL:
http://pages.stern.nyu.edu/~wstarbuc/
The focus is business and management journals, but the discussion is
a lucid and useful analysis of issues and challenges that affect all
fields, especially if the Howard regime controls your life and future
as a scholar.
Andrew van de Ven addresses many of these issues in an important new
book forthcoming on Oxford University Press, Engaged scholarship: A
Guide for Organizational and Social Research.
Visit his personal web site at to read chapter 1 and chapter 9 as PDF
downloads.
http://webpages.csom.umn.edu/smo/avandeven/AHVHOME.htm
Van de Ven discusses the problem of why so much research is not used,
and how we can do better in fields of professional practice.
Gads. I seem to have too much to say these days, but Terry's question
and the further contributions got me rolling.
If the issues of impact and fruitful use interest you, do, please,
have a look at Starbuck and Van de Ven.
Yours,
Ken
--
Prof. Ken Friedman
Institute for Communication, Culture, and Language
Norwegian School of Management
Oslo
Center for Design Research
Denmark's Design School
Copenhagen
+47 46.41.06.76 Tlf NSM
+47 33.40.10.95 Tlf Privat
email: [log in to unmask]
|