Semantic spectrum

From Wikipedia, the free encyclopedia

This article or section needs sources or references that appear in reliable, third-party publications. Alone, primary sources and sources affiliated with the subject of the article are not sufficient for an accurate encyclopedia article. Please include more appropriate citations from reliable sources.
This article has been tagged since February 2007.

The semantic spectrum (sometimes referred to as the ontology spectrum or the smart data continuum or semantic precision) is a series of increasingly precise or rather semantically expressive definitions for data elements in knowledge representations, especially for machine use.

At the low end of the spectrum is a simple binding of a single word or phrase and its definition. At the high end is a full ontology that specifies relationships between data elements using precise URLs for relationships and properties.

With increased specificity comes increased precision and the ability to use tools to automatically integrate systems but also increased cost to build and maintain a metadata registry.

Some steps in the semantic spectrum include the following:

  1. glossary: A simple list of terms and their definitions. A glossary focuses on creating a complete list of the terminology of domain-specific terms and acronyms. It is useful for creating clear and unambiguous definitions for terms and because it can be created with simple word processing tools, few technical tools are necessary.
  2. controlled vocabulary: A simple list of terms, definitions and naming conventions. A controlled vocabulary frequently has some type of oversight process associated with adding or removing data element definitions to ensure consistency. Terms are often defined in relationship to each other.
  3. data dictionary: Terms, definitions, naming conventions and one or more representations of the data elements in a computer system. Data dictionaries often define data types, validation checks such as enumerated values and the formal definitions of each of the enumerated values.
  4. data model: Terms, definitions, naming conventions, representations and one or more representations of the data elements as well as the beginning of specification of the relationships between data elements including abstractions and containers.
  5. taxonomy: A complete data model in an inheritance hierarchy where all data elements inherit their behaviors from a single "super data element". The difference between a data model and a formal taxonomy is the arrangement of data elements into a formal tree structure where each element in the tree is a formally defined concept with associated properties.
  6. ontology: A complete, machine-readable specification of a conceptualization using URLs for all data elements, properties and relationship types. The W3C standard language for representing ontologies is the Web Ontology Language (OWL). Ontologies frequently contain formal business rules formed in discrete logic statements that relate data elements to each another.

Contents

[edit] Choosing a Point on the Semantic Spectrum for your Business

Each organization may choose the positioning for their metadata structures on the semantic spectrum for each organizational unit. The increased cost of movement down the semantic spectrum can be offset by using automated tools to manage the flow of information between computer systems such as databases, workflow or business rule processors. The precise Return on investment (ROI) is difficult to measure since new tools are continually being developed that integrate data models into their core specifications. For example the newly emerging XForms specification now includes a model directly in a web user interface form. Organizations that adopt XForms will see a faster ROI if they build precise data models of their data.

[edit] Typical questions for determining semantic precision

The following is a list of questions that may arise in determining semantic precision.

correctness
How can correct syntax and semantics be enforced? Are tools (such as XML Schema) readily available to validate syntax of data exchanges?
adequacy/expressivity/scope
Does the system represent everything that is of practical use for the purpose? Is an emphasis being placed on data that is externalized (exposed or transferred between systems)?
efficiency
How efficiently can the representation be searched / queried, and - possibly - reasoned on?
complexity
How steep is the learning curve for defining new concepts, querying for them or constraining them? are there appropriate tools for simplifying typical workflows? (See also: ontology editor)
translatability
Can the representation easily be transformed (e.g. by Vocabulary-based transformation) into an equivalent representation so that semantic equivalence is ensured?

[edit] Determining location on the semantic spectrum

Many organizations today are building a metadata registry to store their data definitions and to perform metadata publishing. The question of where they are on the semantic spectrum frequently arises. To determine where your systems are, some of the following questions are frequently useful.

  1. Is there a centralized glossary of terms for the subject matter?
  2. Does the glossary of terms include precise definitions for each terms?
  3. Is there a central repository to store data elements that includes data types information?
  4. Is there an approval process associated with the creation and changes to data elements?
  5. Are coded data elements fully enumerated? Does each enumeration have a full definition?
  6. Is there a process in place to removed duplicate or redundant data elements from the metadata registry?
  7. Is there one or more classification schemes used to classify data elements?
  8. Are document exchanges and web services created using the data elements?
  9. Can the central metadata registry be used as part of a Model-driven architecture?
  10. Are their staff members trained to extract data elements are reuse metadata structures?


[edit] Strategic nature of semantics

Today, much of the world wide web is stored as Hypertext Markup Language. Search engines are severely hampered by their inability to understand the meaning of published web pages. These limitations have created a movement called the Semantic web.

In the past, many organizations that created custom database application used isolated teams of developers that did not formally publish their data definitions. These teams frequently used internal data definitions that were incompatible with other computer systems. This made Enterprise Application Integration and Data warehousing extremely difficult and costly. Many organizations today require that teams consult a centralized data registry before new applications are created.

The job title of an individual that is responsible for coordinating an organization's data is a Data architect.

[edit] References

The first reference to this term was at the AAAI 1999 Ontologies Panel during July 18th and 19th. The panel was organized by Chris Welty, who at the prodding of Fritz Lehmann and in collaboration with the panelists (Fritz, Mike Uschold, Mike Gruninger, and Deborah McGuinness) came up with a "spectrum" of kinds of information systems that were, at the time, referred to as ontologies. The "ontology spectrum" picture appeared in print in the introduction to Formal Ontology and Information Systems: Proceedings of the 2001 Conference ([available from ACM]). The ontology spectrum was also featured in a talk at the Semantics for the Web meeting in 2000 at Dagstuhl by Deborah McGuinness. She wrote up a paper ([[1]]). describing the points on that spectrum that appeared in the book that emerged (much later) from that workshop called "Spinning the Semantic Web" ([[2]]) . Later, Leo Obrst extended the spectrum into two dimensions (which technically is not really a spectrum anymore) and added a lot more detail, which was included in his book, The Semantic Web: A Guide to the Future of XML, Web Services, and Knowledge Management.

The concept of the Semantic precision in business systems was popularized by Dave McComb in his book Semantics in Business Systems : The Savvy Managers Guide published in 2003 where he frequently uses the term Semantic Precision.

This discussion centered around a 10 level partition that included the following levels (listed in the order of increasing semantic precision):

  1. Simple Catalog of Data Elements
  2. Glossary of Terms and Definitions
  3. Thesauri, Narrow Terms, Relationships
  4. Informal "Is-a" relationships
  5. Formal "Is-a" relationships
  6. Formal instances
  7. Frames (properties)
  8. Value Restrictions
  9. Disjointness, Inverse, Part-of
  10. General Logical Constraints

Note that there was a special emphasis on the adding of formal is-a relationships to the spectrum which seems to have been dropped.

The company Cerebra has also popularized this concept by describing the data formats that exist within an enterprise in their ability to store semantically precise metadata. Their list includes:

  1. HTML
  2. PDF
  3. Word Processing documents
  4. Microsoft Excel
  5. Relational databases
  6. XML
  7. XML Schema
  8. Taxonomies
  9. Ontologies

What the concepts share in common is the ability to store information with increasing precision to facilitate intelligent agents.

[edit] See also