Pete Johnston wrote:
> Quoting Madeleine Rothberg <[log in to unmask]>:
>
>> I'm not a DC expert, or even a DC user, so I can't say what the right
>> solutions will be for DC. But yes, a single resource can have multiple
>> access modes adapting distinct access modes in an original. For
>> example, if you have videoA which has audio and visuals, and then
>> videoB, which is the same video with captions and audio descriptions
>> added, then videoB adapts videoA for two different sets of users --
>> those who need alternatives to audio and those who need alternatives
>> to visuals. But perhaps (as was alluded to in Pete's post on "The Real
>> Problem") there can be more than one metadata record and it is up to
>> the system to find them and compare each one to the needs of the user
>> in question, to see if there are useful alternatives available.
>
> Maybe.... but I think an application which aggregated the multiple
> records and merged the descriptions might lose the specificity of
> "B's-text-mode-adapts-A's-visual-mode",
> "B's-audio-mode-adapts-A's-text-mode" (or maybe "adapts" should be "is
> alternative for" in those statements).
>
> It can be done - we just need to be a bit more explicit in the
> descriptions about exactly which things are being related, I think. I'll
> try to put an example together later.
To capture the sort of detail suggested in Madeleine's scenario, I think
you'd have to take an approach something like the attached image i.e.
introduce the notion of "adaptation-processes" or "adaptation-events",
or whatever we want to call them, as resources in their own right -
intermediates between the original resource and the adapted resource, if
you like - each adaptation-process taking exactly one input resource and
output resource, and exactly one input mode and output mode.
Take the names of the properties I used in the diagram with a pinch of
salt, but with something like this, it is made explicit that info
conveyed using the auditory mode in the first resource is conveyed using
the textual mode in the second resource; and info conveyed using the
visual mode in the first resource is conveyed using the auditory mode in
the second resource.
I have to admit this starts to look horribly complex, though!
Pete
--
Pete Johnston
Research Officer (Interoperability)
UKOLN, University of Bath, Bath BA2 7AY, UK
tel: +44 (0)1225 383619 fax: +44 (0)1225 386838
mailto:[log in to unmask]
http://www.ukoln.ac.uk/ukoln/staff/p.johnston/
|