OASIS Mailing List ArchivesView the OASIS mailing list archive below
or browse/search using MarkMail.

 


Help: OASIS Mailing Lists Help | MarkMail Help

 


 

   RE: Soft Landing

[ Lists Home | Date Index | Thread Index ]
  • From: "Bullard, Claude L (Len)" <clbullar@ingr.com>
  • To: KenNorth <KenNorth@email.msn.com>
  • Date: Tue, 24 Oct 2000 08:04:49 -0500

I agree.  The issue is in the overlaps, what Joshua pointed 
out as the "in the sense of" relationships.  At the point where a 
taxonomy needs a magic relationship to include a term (in that case, 
a description probably to further resric the domain), it is likely 
that it now goes extra-domain and into another.  These ecotones are 
important to identify.  It is likely that the rate of change 
here is high and that can be predictive (intuition).  

Domain experts are essential.  Again, there are concerns about  
authority, credentials, testing, and so forth.  The problem of 
automagic is that techniques like frequency of reference can 
be easily poisoned by superstition and adulterate the domains. 
Human scrubbing is required, the domain should be stable, and 
the access rules determined.  There should also be a reasonably 
good payoff because the definitions have to be maintained.

Now, past the economies, how good would you or anyone else 
say that RDF is for this work?   Are techniques or practices 
emerging for building RDFs similar to those described in 
the Principia Cybernetica or perhaps as Roger Costello has 
been assembling?  There is undoubted power to having a single 
language like RDF for doing this work and that power emerges 
as the domains become stable and interrelated.  I should expect 
the portal builders to be able to take enormous advantage of 
such a polyglot for resource discovery.  Beyond that, in the 
area TimBL describes for business, the routing of application 
messages to such services seems useful on the surface, but I 
have a wait and see attitude.  Coherence, transparency, stability 
and reliability have to be assessed and I don't know how that 
is done yet.  

Len Bullard
Intergraph Public Safety
clbullar@ingr.com
http://www.mp3.com/LenBullard

Ekam sat.h, Vipraah bahudhaa vadanti.
Daamyata. Datta. Dayadhvam.h


-----Original Message-----
From: KenNorth [mailto:KenNorth@email.msn.com]

If I were king of the world, with unlimited budget and unlimited
cooperation, I'd start with a taxonomy and domain experts. Let them define a
domain vocabulary (again I keep pointing to MeSH for medical literature).

Then, when new literature is published each month, run it through machine
analysis to identify new terms that start popping up in the literature
(e.g., XML a few years ago). Also identify relationships to existing
concepts or terms (similarity searches), and so on.  The domain experts
identify an alert level (e.g., 5 citations) and when a term or concept
exceeds that level, it's included in a monthly update they receive -- new
terms and concepts in the literature. They use that information when
updating domain vocabularies on a quarterly basis.

Using a pre-defined domain vocabulary is probably more efficient than doing
it all automagically using inference engines, machine analysis of schemas,
RDF, parsing and so on.

Look at the portals that migrated to a classification scheme, instead of
being simply keyword container searches.













 

News | XML in Industry | Calendar | XML Registry
Marketplace | Resources | MyXML.org | Sponsors | Privacy Statement

Copyright 2001 XML.org. This site is hosted by OASIS