Print

Print


Hi Bridget, I have been mulling over your thoughts during Christmas (by
way of pleasant relief from family gatherings and rubbish TV ;-).  I
agree that it's harder to measure the success of your strategies if the
learning outcomes are loosely expressed.  But it seems to me that a goal
such as encouraging open-ended discovery need not necessarily result in
loosely expressed outcomes.  So, taking your example "develop skills in
comparison of objects", as designers we might usefully think about
comparison skills: what are they?  How are they developed?  What are the
common barriers or pitfalls to developing such skills?  Then maybe we
could begin to develop a learning activity which encourages people to
make comparisons and which helps them to develop comparative skills.
This would be an open ended activity in the sense that the learners
would be free to choose what they compared but the learning outcome is
quite specific and quite easy to measure therefore.   So I am not sure
that the opposite of measurable learning activities is open ended
discovery.  Putting that another way, I think it is possible to define
specific learning outcomes for open ended discovery activities providing
one accepts a shift in focus from learning content to the learning
process.  For example, we might devise a learning activity in which
learning comparison skills in the context of medieval surgical
implements is more important than learning about medieval surgical
implements themselves.  Such an activity would be quite exploratory and
open ended in one sense and yet focussed and precise in another.

A consequence of loosely expressed learning outcomes is that one has to
accept that the actual learning emerges from the interaction between
learner and learning experience without being able to predict it very
precisely in advance. But from a design perspective this approach has a
major flaw.  Under this regime a good learning activity design will be
one that engenders a lot of positive responses.  But until the design
has undergone extensive testing, with large numbers of participants, its
overall impact on users will not be apparent. There will not be enough
interactions for the overall pattern to emerge.  Given the tight
budgetary and time constraints that surround most web development
projects, lengthy field trials are usually out of the question during
the development phase.  This effectively leaves the designer without a
tool for testing the effectiveness of the design until most of the
investment has been made.

So while "emergent" learning outcomes are useful for retrospective
assessment of overall impact I think they are of little use for
supporting a developmental user-centred design approach.  We need
"predictive" measures to assess performance iteratively against clearly
specified intended learning outcomes during the design development
process.  I hope the above exmple explins why I think that clearly
specified learning outcomes need not be didactic.. I think a couple of
good real life examples are:  "AccessArt"
(http://www.accessart.org.uk/index.php), an online learning resource for
"pupils, students, and lifelong learners as well as teachers, gallery
educators and artists" is a good example of how online resources can be
used to support clear learning objectives without prescribing the
precise outcomes.  The online drawing workshop for 16+ learners
encourages and supports exploration of understanding of seeing and
drawing.  Another good example is the set of learning "challenges" in
the British Museum's Ancient Egypt Interactive learning web site (see
for example the temple challenge at
http://www.ancientegypt.co.uk/temples/home.html.  In both cases the
activities provided carefully "scaffold" the learning experience.  

Id be interested in responses to these ideas and suggestions of further
examples.

Stephen
-----Original Message-----
From: Museums Computer Group [mailto:[log in to unmask]] On Behalf Of
Bridget McKenzie
Sent: 01 December 2006 11:16
To: [log in to unmask]
Subject: Re: Learning design


Hello
This is an interjection, I hope a little bit relevant, just to raise a 
thought about instructional learning design in a context that favours 
constructivist learning.

Museum learning is about open-ended discovery, so learning outcomes will

often be loosely expressed e.g. 'develop skills in comparison of
objects', 
'develop tolerance of other cultures'. Therefore it's harder to measure
the 
success of your strategies. You may be able to discern changes in
learners 
if you evaluate a sustained project, including museum visits,
experimental 
activities and use of the linked web resources. Or formative evaluation
can 
work if the resources are seen being used in a discovery learning
context. 
If you haven't got time/money to do this you can only guess that
open-ended 
tools would be successful in reaching the desired outcomes if people
seem to 
like them and are using them.

It could be tempting to design more and more instructional or didactic
web 
tools, simply because you know you're supposed to evaluate them and it's

easier to assess whether users 'got it' or not.

Much better to support an expansion of e-learning as dialogue, so that 
discovery happens online in interactions between people with questions, 
people with insights and cultural artefacts. The proof of that kind of 
pudding is definitely in the eating, and not only divined in an
expensive 
evaluation.

Bridget

----- Original Message ----- 
From: "Joe Cutting" <[log in to unmask]>
To: <[log in to unmask]>
Sent: Friday, December 01, 2006 10:34 AM
Subject: Learning design


> Stephen,
> >>
> Hi, does anyone have experience of using user-centred design to create

> either exhibitions/displays and/or web sites please? I would be really

> interested to find out how useful you found this approach.
> >>
> Dear colleagues, have you built learning activities into your Museum 
> web site?  Was it important for you to be able to demonstrate the 
> effectiveness of your designs?  Did you test the designs during 
> development, and/or after they were completed?
> >>
>
> This is pretty much established best practice and becoming more 
> widespread
> as
> funders are demanding summative evaluation on exhibitions and projects

> after completion.
> They're also demanding that institutions specify objectives and
audiences 
> upfront with their
> funding applications - although I would agree that these can easily
get 
> lost along the way.
> It tends to be much easier to do formative evaluation on exhibits and 
> micro-sites than
> whole institution sites due to the difficulty of specifying audiences
and 
> objectives for a whole institution site.
> The amount of evaluation done tends to depend on the attitudes and 
> resources of the institution but if you're looking into this
> your main issues are going to be
> Institutions don't tend to release the results of evaluation unless 
> they're completely positive. This is for a variety of political
> reasons connected with the press and funders.
> There's also very little movement by institutions to publish anything 
> about their development methods and
> experiences - in general its just not seen as a priority. There are 
> honourable exceptions like the Tate's Multimedia tour page
> http://www.tate.org.uk/modern/multimediatour/re_keyfindings.htm.
There's 
> also bits and pieces by Ben Gammon scattered around the web
> like this one
http://ukupa.org.uk/events/presentations/science_museum.pdf 
> (google "Ben Gammon" for more).
>
> Given that there's a lot of work going on, your problem is unlikely to

> be
> finding examples - its much more likely to be
> getting people to tell you about them in any detail - particularly if
user 
> testing wasn't done or produced "bad" results.
> My recommendation would be either to get hold of a list of recent
projects 
> from a major funder like HLF or Wellcome and then
> go and interview a selection of the project managers or pick one
project 
> which did use a lot of user testing and study it in detail.
> To give you an idea, one project I worked on last year did around 20 
> formative evaluation studies in 4 months of development so there's a
> lot to get your teeth into.
>
> All the best with your project
>
> Joe
>
>
>
>
> Joe Cutting
> Computer exhibits and installations
> www.joecutting.com
> The Fishergate Centre, 4 Fishergate, York, YO10 4FB
> 01904 624681
>
> As of 30th October 2006 I have a new office so
> please note my new address and phone number
> **************************************************
> For mcg information and to manage your subscription to the list, visit

> the
> website at http://www.museumscomputergroup.org.uk
> **************************************************
>
> 

**************************************************
For mcg information and to manage your subscription to the list, visit
the website at http://www.museumscomputergroup.org.uk
**************************************************

**************************************************
For mcg information and to manage your subscription to the list, visit the website at http://www.museumscomputergroup.org.uk
**************************************************