Dependency relation

For other uses, see Dependency (disambiguation).
Not to be confused with Dependence relation, which is a generalization of the concept of linear dependence among members of a vector space.

In mathematics and computer science, a dependency relation is a binary relation that is finite, symmetric, and reflexive; i.e. a finite tolerance relation. That is, it is a finite set of ordered pairs D, such that

In general, dependency relations are not transitive; thus, they generalize the notion of an equivalence relation by discarding transitivity.

Let \Sigma denote the alphabet of all the letters of D. Then the independency induced by D is the binary relation I

I = \Sigma \times \Sigma \setminus D

That is, the independency is the set of all ordered pairs that are not in D. The independency is symmetric and irreflexive.

The pairs (\Sigma, D) and (\Sigma, I), or the triple (\Sigma, D, I) (with I induced by D) are sometimes called the concurrent alphabet or the reliance alphabet.

The pairs of letters in an independency relation induce an equivalence relation on the free monoid of all possible strings of finite length. The elements of the equivalence classes induced by the independency are called traces, and are studied in trace theory.

Examples

Consider the alphabet \Sigma=\{a,b,c\}. A possible dependency relation is


\begin{align} D &= \{a,b\}\times\{a,b\} \quad \cup \quad \{a,c\}\times\{a,c\} \\
 &= \{a,b\}^2 \cup \{a,c\}^2 \\
 &= \{ (a,b),(b,a),(a,c),(c,a),(a,a),(b,b),(c,c)\} 
\end{align}

The corresponding independency is

I_D=\{(b,c)\,,\,(c,b)\}

Therefore, the letters b,c commute, or are independent of one another.

This article is issued from Wikipedia - version of the Tuesday, June 23, 2015. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.