Semantic spectrum

Semantic spectrum

The semantic spectrum (sometimes referred to as the ontology spectrum or the smart data continuum or semantic precision) is a series of increasingly precise or rather semantically expressive definitions for data elements in knowledge representations, especially for machine use.

At the low end of the spectrum is a simple binding of a single word or phrase and its definition. At the high end is a full ontology that specifies relationships between data elements using precise URIs for relationships and properties.

With increased specificity comes increased precision and the ability to use tools to automatically integrate systems but also increased cost to build and maintain a metadata registry.

Some steps in the semantic spectrum include the following:
# glossary: A simple list of terms and their definitions. A glossary focuses on creating a complete list of the terminology of domain-specific terms and acronyms. It is useful for creating clear and unambiguous definitions for terms and because it can be created with simple word processing tools, few technical tools are necessary.
# controlled vocabulary: A simple list of terms, definitions and naming conventions. A controlled vocabulary frequently has some type of oversight process associated with adding or removing data element definitions to ensure consistency. Terms are often defined in relationship to each other.
# data dictionary: Terms, definitions, naming conventions and one or more representations of the data elements in a computer system. Data dictionaries often define data types, validation checks such as enumerated values and the formal definitions of each of the enumerated values.
# data model: Terms, definitions, naming conventions, representations and one or more representations of the data elements as well as the beginning of specification of the relationships between data elements including abstractions and containers.
# taxonomy: A complete data model in an inheritance hierarchy where all data elements inherit their behaviors from a single "super data element". The difference between a data model and a formal taxonomy is the arrangement of data elements into a formal tree structure where each element in the tree is a formally defined concept with associated properties.
# ontology: A complete, machine-readable specification of a conceptualization using URLs for all data elements, properties and relationship types. The W3C standard language for representing ontologies is the Web Ontology Language (OWL). Ontologies frequently contain formal business rules formed in discrete logic statements that relate data elements to each another.

Choosing a Point on the Semantic Spectrum for your Business

Each organization may choose the positioning for their metadata structures on the semantic spectrum for each organizational unit. The increased cost of movement down the semantic spectrum can be offset by using automated tools to manage the flow of information between computer systems such as databases, workflow or business rule processors. The precise Return on investment (ROI) is difficult to measure since new tools are continually being developed that integrate data models into their core specifications. For example the newly emerging XForms specification now includes a model directly in a web user interface form. Organizations that adopt XForms will see a faster ROI if they build precise data models of their data.

Typical questions for determining semantic precision

The following is a list of questions that may arise in determining semantic precision.

;correctness: How can correct syntax and semantics be enforced? Are tools (such as XML Schema) readily available to validate syntax of data exchanges?;adequacy/expressivity/scope: Does the system represent everything that is of practical use for the purpose? Is an emphasis being placed on data that is externalized (exposed or transferred between systems)?;efficiency: How efficiently can the representation be searched / queried, and - possibly - reasoned on?;complexity: How steep is the learning curve for defining new concepts, querying for them or constraining them? are there appropriate tools for simplifying typical workflows? (See also: ontology editor);translatability: Can the representation easily be transformed (e.g. by Vocabulary-based transformation) into an equivalent representation so that semantic equivalence is ensured?

Determining location on the semantic spectrum

Many organizations today are building a metadata registry to store their data definitions and to perform metadata publishing. The question of where they are on the semantic spectrum frequently arises. To determine where your systems are, some of the following questions are frequently useful.

# Is there a centralized glossary of terms for the subject matter?
# Does the glossary of terms include precise definitions for each terms?
# Is there a central repository to store data elements that includes data types information?
# Is there an approval process associated with the creation and changes to data elements?
# Are coded data elements fully enumerated? Does each enumeration have a full definition?
# Is there a process in place to removed duplicate or redundant data elements from the metadata registry?
# Is there one or more classification schemes used to classify data elements?
# Are document exchanges and web services created using the data elements?
# Can the central metadata registry be used as part of a Model-driven architecture?
# Are their staff members trained to extract data elements are reuse metadata structures?

trategic nature of semantics

Today, much of the world wide web is stored as Hypertext Markup Language. Search engines are severely hampered by their inability to understand the meaning of published web pages. These limitations have created a movement called the Semantic web.

In the past, many organizations that created custom database application used isolated teams of developers that did not formally publish their data definitions. These teams frequently used internal data definitions that were incompatible with other computer systems. This made Enterprise Application Integration and Data warehousing extremely difficult and costly. Many organizations today require that teams consult a centralized data registry before new applications are created.

The job title of an individual that is responsible for coordinating an organization's data is a Data architect.

History

The first reference to this term was at the [http://www.aaai.org/ AAAI] 1999 [http://www.cs.vassar.edu/faculty/welty/presentations/aaai-99/ Ontologies Panel] during July 18th and 19th. The panel was organized by Chris Welty, who at the prodding of Fritz Lehmann andin collaboration with the panelists (Fritz, Mike Uschold, Mike Gruninger, and Deborah McGuinness) came up with a "spectrum" of kinds of information systems that were, at the time, referred to as ontologies. The "ontology spectrum" picture appeared in print in the introduction to"Formal Ontology and Information Systems: Proceedings of the 2001 Conference" ( [http://portal.acm.org/toc.cfm?id=505168&coll=GUIDE&dl=GUIDE&type=proceeding&CFID=10676962&CFTOKEN=42478708 available from ACM] ). The ontology spectrum was also featured in a talk at the Semantics for the Web meeting in 2000 at Dagstuhl by Deborah McGuinness. She wrote up a paper ( [http://www.ksl.stanford.edu/people/dlm/papers/ontologies-come-of-age-mit-press-(with-citation).htm] ). describing the points on that spectrum that appeared in the book that emerged (much later) from that workshop called "Spinning the Semantic Web" ( [http://mitpress.mit.edu/catalog/item/default.asp?ttype=2&tid=9182] ) . Later, Leo Obrst extended the spectrum into two dimensions (which technically is not really a spectrum anymore) and added a lot more detail, which was included in his book, "The Semantic Web: A Guide to the Future of XML, Web Services, and Knowledge Management".

The concept of the Semantic precision in business systems was popularized by Dave McComb in his book "Semantics in Business Systems : The Savvy Managers Guide" published in 2003 where he frequently uses the term Semantic Precision.

This discussion centered around a 10 level partition that included the following levels (listed in the order of increasing semantic precision):
# Simple Catalog of Data Elements
# Glossary of Terms and Definitions
# Thesauri, Narrow Terms, Relationships
# Informal "Is-a" relationships
# Formal "Is-a" relationships
# Formal instances
# Frames (properties)
# Value Restrictions
# Disjointness, Inverse, Part-of
# General Logical Constraints

Note that there was a special emphasis on the adding of formal "is-a" relationships to the spectrum which seems to have been dropped.

The company [http://cerebra.com Cerebra] has also popularized this concept by describing the data formats that exist within an enterprise in their ability to store semantically precise metadata. Their list includes:
# HTML
# PDF
# Word Processing documents
# Microsoft Excel
# Relational databases
# XML
# XML Schema
# Taxonomies
# Ontologies

What the concepts share in common is the ability to store information with increasing precision to facilitate intelligent agents.

See also

* Enterprise messaging system
* Semantics
* SKOS
* ontology (computer science)
* collabulary
* Web service
* Conceptual Interoperability

References

* "Semantics in Business Systems: The Savvy Managers Guide", Dave McComb, 2003


Wikimedia Foundation. 2010.

Игры ⚽ Поможем написать реферат

Look at other dictionaries:

  • Semantic interoperability — (also referred to as Computable Semantic Interoperability) is the ability of two or more computer systems to exchange information and have the meaning of that information automatically interpreted by the receiving system accurately enough to… …   Wikipedia

  • Semantics — is the study of meaning in communication. The word derives from Greek σημαντικός ( semantikos ), significant , [cite web|url=http://www.perseus.tufts.edu/cgi bin/ptext?doc=Perseus%3Atext%3A1999.04.0057%3Aentry%3D%2393797|title=Semantikos, Henry… …   Wikipedia

  • Knowledge representation — is an area in artificial intelligence that is concerned with how to formally think , that is, how to use a symbol system to represent a domain of discourse that which can be talked about, along with functions that may or may not be within the… …   Wikipedia

  • Controlled vocabulary — Controlled vocabularies provide a way to organize knowledge for subsequent retrieval. They are used in subject indexing schemes, subject headings, thesauri, taxonomies and other form of knowledge organization systems. Controlled vocabulary… …   Wikipedia

  • Vocabulary-based transformation — In metadata, a vocabulary based transformation (VBT) is a transformation aided by the use of a semantic equivalence statements within a controlled vocabulary.Many organizations today require communication between one or more computers. Although… …   Wikipedia

  • Представление знаний — Представление знаний  вопрос, возникающий в когнитологии (науке о мышлении), в информатике и в исследованиях искусственного интеллекта. В когнитологии он связан с тем, как люди хранят и обрабатывают информацию. В информатике  с подбором …   Википедия

  • Представление информации — Представление знаний  вопрос, возникающий в когнитологии (науке о мышлении), в информатике и в искусственном интеллекте. В когнитологии он связан с тем, как люди хранят и обрабатывают информацию. В информатике  основная цель  подбор представления …   Википедия

  • Meta-Process Modeling — is a type of metamodeling used in software engineering and systems engineering for the analysis and construction of models applicable and useful some predefined problems. Meta process support the effort of creating flexible process models. The… …   Wikipedia

  • Data Reference Model — The DRM Collaboration Process. The Data Reference Model (DRM) is one of the five reference models of the Federal Enterprise Architecture (FEA). Contents 1 Overview …   Wikipedia

  • Glossary of Unified Modeling Language terms — This glossary of Unified Modeling Language terms covers all versions of UML. Individual entries will point out any distinctions that exist between versions.A* Abstract An indicator applied to a classifier (e.g., actor, class, use case) or to some …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”