Dependency relation
From Wikipedia, the free encyclopedia
This article does not cite any references or sources. (March 2008) Please help improve this article by adding citations to reliable sources. Unverifiable material may be challenged and removed. |
In mathematics and computer science, a dependency relation is a binary relation that is finite, symmetric, and reflexive. That is, it is a finite set of ordered pairs D, such that
- If then (symmetric)
- If a is an element of the set on which the relation is defined, then (reflexive)
In general, dependency relations are not transitive; thus, they generalize the notion of an equivalence relation by discarding transitivity.
Let Σ denote the alphabet of all the letters of D. Then the independency induced by D is the binary relation I
That is, the independency is the set of all ordered pairs that are not in D. Clearly, the independency is symmetric and irreflexive.
The pairs (Σ,D) and (Σ,I), or the triple (Σ,D,I) (with I induced by D) are sometimes called the concurrent alphabet or the reliance alphabet.
The pairs of letters in an independency relation induce an equivalence relation on the free monoid of all possible strings of finite length. The elements of the equivalence classes induced by the independency are called traces, and are studied in trace theory.
[edit] Examples
Consider the alphabet Σ = {a,b,c}. A possible dependency relation is
The corresponding independency is
Therefore, the letters b,c commute, or are independent of one-another.