[log in to unmask]" type="cite">Yeah, I agree completely. There are plenty of hacks out there if all you want is a black box (google for "statistical parsing"). What I want is a glass box -- that's the whole point of computer simulation, especially when the model is supposed to be informed by theory. We've begun exploring some techniques for transforming an underlying network data structure into something simpler to visualize in the user interface of the simulator (e.g. each property arc uses a singleton node to represent the relation, and these all have to be distinguished internally with something like a serial number -- but you don't see those numbers except in debugging mode). Obviously, there's no limit to the complexity of the transformation between the actual network being simulated and the view of it presented in the user interface -- as long as we're always straight on what's really there underneath. So my vote, as you predicted, would be to make the linguistic representations as complex as they need to be in order to capture the phenomena. The more mechanical bits (as in your hack-free solution, which I like) can be abstracted out when we don't need to see them. -- Mark jasper holmes wrote:So the question is, if both of these solutions can be made to work, which should we prefer. The answer to that is, of course, that it depends what you want it for. You might take the view (I do) that the second is the most psychologically real. But then you might feel at one and the same time that the first would be easier to implement in an algorithm, especially if you were already using [x before y] and [x after y] in your computer application to save hassle (and I know it does save hassle: when you start counting every little arc they soon add up). You may or may not (I do) want to bear in mind that this is and remains a hack. So what do you want? Psychological reality or a working computer model? Mark, I guess, would argue that if your hacks don't correspond to psychological reality then sooner or later you are going to come up against something that you can't model any more without going back and undoing the hacks. Dick might also say that the whole point of the computer model is to represent what we really think grammar is like (to test our theory of grammar as much as anything else). The principles of WG plus the correct grammar structures should always give the correct result; and if they don't then you need to revisit the grammar (or the principles!). Japs On 9/4/08, jasper holmes <[log in to unmask]> wrote:I don't mean by cutting in at the beginning like this to discount the very useful discussion that has already taken place. It's just I couldn't find anywhere else to sensibly come in. This is a tricky one, isn't it and I remember we struggled with it when we were in the same situation before. I think, though, that we largely solved it, or caused it to disappear. I think we looked at possible ways of putting two relationships into conflict without one (transitive)isa-ing the other, rather as Lyne has been suggesting, so that 13 would block 14 because of 'after' overriding 'before'. We tried to state this in terms of some common ancestor, as Dick is doing below. I think my favourite solution along these lines was to say that there are some relationships that you can only have one of: you can have as many dependents as you like so [X extractee Y] doesn't override [X object Y], but you can only have one 'ordering' (let's call it that for now, until the next paragraph anyway), so [X after Y] does override [X before Y]. Can't remember exactly why this didn't work out, but I don't think it did. Fortunately, however, there is another solution, and this follows Lynes other suggestion: [X after Y] and [X before Y] are not the right way to represent the ordering relationships. I'd say something more like this: [91: word dependent X] [92: word time Pw] [93: X time PX] [94: Pw < PX] ('less than') [95: subject isa dependent] [96: word subject Y] [97: Y time PY] [98: Pw > PY] ('greater than') Then, for what it's worth, 97 overrides 93, since Y isa X. I've been trying to think of other examples, in case they aren't amenable to this kind of solution, but I can't. Perhaps there are some; it's a bit of a hostage to fortune I guess to rely on none turning up. I think I've presented two almost solutions. I evaluate them (very briefly) in the next message. Then there follows a message with a red herring. Jasper On 8/28/08, Richard Hudson <[log in to unmask]> wrote: > > Dear All, > After a long silence on this list, here's a question for you all. It's > about how to make default inheritance work properly. We (at least, Mark Line > and I) have an algorithm which promises to work reasonably smoothly in a > computer system that Mark is building (and that should work more generally > as well, of course), but only for one of the two kinds of situation that > default inheritance has to deal with. My question is whether anyone has any > bright ideas for handling the other kind. Here goes with the problem. > > DI has to take as input a proposition [1: A R V], where A is the argument, > R is the relation and V is the value, and apply it to some instance of A, > called A'. "Applying it" means deciding whether or not to inherit [2:A' R' > V'], a copy of [1], in the light of the store of propositions P already > stored for A'; and the crucial question is whether P contains a proposition > which overrides [2]. > > For example, assume this database: > [3: Bird locomotion flying] > [4: Penguin locomotion swimming] > [5: Penguin is-a bird] > > Store of propositions about Penguin', some particular penguin: > [6: Penguin' isa Penguin] > [4': Penguin' locomotion' swimming'] {inherited from [4]}, where > [locomotion' is-a locomotion] > > Question: can Penguin' also inherit [3']? > [3': Penguin' locomotion' flying'] > > The question assumes a potentially inheritable stored proposition IP and > some potential overriding proposition OP - e.g. in the above IP = [3] and OP > = [4']. > > Type 1 inheritance: > Where IP and OP have the same relation but different values. This is easy, > because we can define 'the same relation' as being where: > > IP = [A1 R1 V1] > OP = [A2 R2 V2] > and [R2 is-a R1]. > There's not even any need to check the relation between V1 and V2, because > it doesn't matter whether or not they're related; either way, the > inheritance system ignores IP. > > Type 2 inheritance: > Where IP and OP have the same value but different relations. This is the > hard one, and I'm embarrassed to say that although I've been aware of the > problem for years, I've also managed to avoid thinking about it. It's > painfully easy to illustrate from word order rules: > [7: word dependent X] > [8: word before X] > [9: word subject Y] > [10: subject is-a dependent] > [11: word after Y] > > Precisely what is it that prevents some word W from inheriting the > following? > [12: W subject Z] > [13: W after Z] > [14: W before Z] > I've had various thoughts, but none that I really like, so I'd be > interested to hear other ideas. > > Best wishes, Dick > > > > -- > > > Richard Hudson, FBA. Emeritus Professor, University College London > > My web page: www.phon.ucl.ac.uk/home/dick/home.htm > Why I support the academic boycott of Israel: > www.phon.ucl.ac.uk/home/dick/home.htm#boycott > My latest book: Language Networks. The New Word Grammar > >-- Mark Mark P. Line Bartlesville, OK
Richard Hudson, FBA. Emeritus Professor, University College London