Trace monoid

Generalization of strings in computer science

In computer science, a trace is an equivalence class of strings, wherein certain letters in the string are allowed to commute, but others are not. Traces generalize the concept of strings by relaxing the requirement for all the letters to have a definite order, instead allowing for indefinite orderings in which certain reshufflings could take place. In an opposite way, traces generalize the concept of sets with multiplicities by allowing for specifying some incomplete ordering of the letters rather than requiring complete equivalence under all reorderings. The trace monoid or free partially commutative monoid is a monoid of traces.

Traces were introduced by Pierre Cartier and Dominique Foata in 1969 to give a combinatorial proof of MacMahon's master theorem. Traces are used in theories of concurrent computation, where commuting letters stand for portions of a job that can execute independently of one another, while non-commuting letters stand for locks, synchronization points or thread joins.[1]

The trace monoid is constructed from the free monoid (the set of all strings of finite length) as follows. First, sets of commuting letters are given by an independency relation. These induce an equivalence relation of equivalent strings; the elements of the equivalence classes are the traces. The equivalence relation then partitions the elements of the free monoid into a set of equivalence classes; the result is still a monoid; it is a quotient monoid now called the trace monoid. The trace monoid is universal, in that all dependency-homomorphic (see below) monoids are in fact isomorphic.

Trace monoids are commonly used to model concurrent computation, forming the foundation for process calculi. They are the object of study in trace theory. The utility of trace monoids comes from the fact that they are isomorphic to the monoid of dependency graphs; thus allowing algebraic techniques to be applied to graphs, and vice versa. They are also isomorphic to history monoids, which model the history of computation of individual processes in the context of all scheduled processes on one or more computers.

Trace

Let Σ {\displaystyle \Sigma ^{*}} denote the free monoid on a set of generators Σ {\displaystyle \Sigma } , that is, the set of all strings written in the alphabet Σ {\displaystyle \Sigma } . The asterisk is a standard notation for the Kleene star. An independency relation I {\displaystyle I} on the alphabet Σ {\displaystyle \Sigma } then induces a symmetric binary relation {\displaystyle \sim } on the set of strings Σ {\displaystyle \Sigma ^{*}} : two strings u , v {\displaystyle u,v} are related, u v , {\displaystyle u\sim v,} if and only if there exist x , y Σ {\displaystyle x,y\in \Sigma ^{*}} , and a pair ( a , b ) I {\displaystyle (a,b)\in I} such that u = x a b y {\displaystyle u=xaby} and v = x b a y {\displaystyle v=xbay} . Here, u , v , x {\displaystyle u,v,x} and y {\displaystyle y} are understood to be strings (elements of Σ {\displaystyle \Sigma ^{*}} ), while a {\displaystyle a} and b {\displaystyle b} are letters (elements of Σ {\displaystyle \Sigma } ).

The trace is defined as the reflexive transitive closure of {\displaystyle \sim } . The trace is thus an equivalence relation on Σ {\displaystyle \Sigma ^{*}} and is denoted by D {\displaystyle \equiv _{D}} , where D {\displaystyle D} is the dependency relation corresponding to I . {\displaystyle I.} D = ( Σ × Σ ) I {\displaystyle D=(\Sigma \times \Sigma )\setminus I} and I = ( Σ × Σ ) D . {\displaystyle I=(\Sigma \times \Sigma )\setminus D.} Different independencies or dependencies will give different equivalence relations.

The transitive closure implies that u D v {\displaystyle u\equiv _{D}v} if and only if there exists a sequence of strings ( w 0 , w 1 , , w n ) {\displaystyle (w_{0},w_{1},\cdots ,w_{n})} such that u w 0 , {\displaystyle u\sim w_{0},} v w n , {\displaystyle v\sim w_{n},} and w i w i + 1 {\displaystyle w_{i}\sim w_{i+1}} for all 0 i < n {\displaystyle 0\leq i<n} . The trace is stable under the monoid operation on Σ {\displaystyle \Sigma ^{*}} , i.e., concatenation, and D {\displaystyle \equiv _{D}} is therefore a congruence relation on Σ . {\displaystyle \Sigma ^{*}.}

The trace monoid, commonly denoted as M ( D ) {\displaystyle \mathbb {M} (D)} , is defined as the quotient monoid

M ( D ) = Σ / D . {\displaystyle \mathbb {M} (D)=\Sigma ^{*}/\equiv _{D}.}

The homomorphism

ϕ D : Σ M ( D ) {\displaystyle \phi _{D}:\Sigma ^{*}\to \mathbb {M} (D)}

is commonly referred to as the natural homomorphism or canonical homomorphism. That the terms natural or canonical are deserved follows from the fact that this morphism embodies a universal property, as discussed in a later section.

One will also find the trace monoid denoted as M ( Σ , I ) {\displaystyle M(\Sigma ,I)} where I {\displaystyle I} is the independency relation. One can also find the commutation relation used instead of the independency relation; it differs from the independency relation by also including all the diagonal elements of Σ × Σ {\textstyle \Sigma \times \Sigma } since letters "commute with themselves" in a free monoid of strings of those letters.

Examples

Consider the alphabet Σ = { a , b , c } {\displaystyle \Sigma =\{a,b,c\}} . A possible dependency relation is

D = { a , b } × { a , b } { a , c } × { a , c } = { a , b } 2 { a , c } 2 = { ( a , b ) , ( b , a ) , ( a , c ) , ( c , a ) , ( a , a ) , ( b , b ) , ( c , c ) } . {\displaystyle {\begin{matrix}D&=&\{a,b\}\times \{a,b\}\quad \cup \quad \{a,c\}\times \{a,c\}\\&=&\{a,b\}^{2}\cup \{a,c\}^{2}\\&=&\{(a,b),(b,a),(a,c),(c,a),(a,a),(b,b),(c,c)\}.\end{matrix}}}

The corresponding independency is

I D = { ( b , c ) , ( c , b ) } . {\displaystyle I_{D}=\{(b,c)\,,\,(c,b)\}.}

Therefore, the letters b , c {\displaystyle b,c} commute. Thus, for example, a trace equivalence class for the string a b a b a b b c a {\displaystyle abababbca} would be

[ a b a b a b b c a ] D = { a b a b a b b c a , a b a b a b c b a , a b a b a c b b a } {\displaystyle [abababbca]_{D}=\{abababbca\,,\;abababcba\,,\;ababacbba\}}

and the equivalence class [ a b a b a b b c a ] D {\displaystyle [abababbca]_{D}} would be an element of the trace monoid.

Properties

The cancellation property states that equivalence is maintained under right cancellation. That is, if w v {\displaystyle w\equiv v} , then ( w ÷ a ) ( v ÷ a ) {\displaystyle (w\div a)\equiv (v\div a)} . Here, the notation w ÷ a {\displaystyle w\div a} denotes right cancellation, the removal of the first occurrence of the letter a from the string w, starting from the right-hand side. Equivalence is also maintained by left-cancellation. Several corollaries follow:

  • Embedding: w v {\displaystyle w\equiv v} if and only if x w y x v y {\displaystyle xwy\equiv xvy} for strings x and y. Thus, the trace monoid is a syntactic monoid.[non sequitur See Talk:Trace monoid#As Syntactic Monoids]
  • Independence: if u a v b {\displaystyle ua\equiv vb} and a b {\displaystyle a\neq b} , then a is independent of b. That is, ( a , b ) I D {\displaystyle (a,b)\in I_{D}} . Furthermore, there exists a string w such that u w b {\displaystyle u\equiv wb} and v w a {\displaystyle v\equiv wa} .
  • Projection rule: equivalence is maintained under string projection, so that if w v {\displaystyle w\equiv v} , then π Σ ( w ) π Σ ( v ) {\displaystyle \pi _{\Sigma }(w)\equiv \pi _{\Sigma }(v)} .

A strong form of Levi's lemma holds for traces. Specifically, if u v x y {\displaystyle uv\equiv xy} for strings u, v, x, y, then there exist strings z 1 , z 2 , z 3 {\displaystyle z_{1},z_{2},z_{3}} and z 4 {\displaystyle z_{4}} such that ( w 2 , w 3 ) I D {\displaystyle (w_{2},w_{3})\in I_{D}} for all letters w 2 Σ {\displaystyle w_{2}\in \Sigma } and w 3 Σ {\displaystyle w_{3}\in \Sigma } such that w 2 {\displaystyle w_{2}} occurs in z 2 {\displaystyle z_{2}} and w 3 {\displaystyle w_{3}} occurs in z 3 {\displaystyle z_{3}} , and

u z 1 z 2 , v z 3 z 4 , {\displaystyle u\equiv z_{1}z_{2},\qquad v\equiv z_{3}z_{4},}
x z 1 z 3 , y z 2 z 4 . {\displaystyle x\equiv z_{1}z_{3},\qquad y\equiv z_{2}z_{4}.} [2]

Universal property

A dependency morphism (with respect to a dependency D) is a morphism

ψ : Σ M {\displaystyle \psi :\Sigma ^{*}\to M}

to some monoid M, such that the "usual" trace properties hold, namely:

1. ψ ( w ) = ψ ( ε ) {\displaystyle \psi (w)=\psi (\varepsilon )} implies that w = ε {\displaystyle w=\varepsilon }
2. ( a , b ) I D {\displaystyle (a,b)\in I_{D}} implies that ψ ( a b ) = ψ ( b a ) {\displaystyle \psi (ab)=\psi (ba)}
3. ψ ( u a ) = ψ ( v ) {\displaystyle \psi (ua)=\psi (v)} implies that ψ ( u ) = ψ ( v ÷ a ) {\displaystyle \psi (u)=\psi (v\div a)}
4. ψ ( u a ) = ψ ( v b ) {\displaystyle \psi (ua)=\psi (vb)} and a b {\displaystyle a\neq b} imply that ( a , b ) I D {\displaystyle (a,b)\in I_{D}}

Dependency morphisms are universal, in the sense that for a given, fixed dependency D, if ψ : Σ M {\displaystyle \psi :\Sigma ^{*}\to M} is a dependency morphism to a monoid M, then M is isomorphic to the trace monoid M ( D ) {\displaystyle \mathbb {M} (D)} . In particular, the natural homomorphism is a dependency morphism.

Normal forms

There are two well-known normal forms for words in trace monoids. One is the lexicographic normal form, due to Anatolij V. Anisimov and Donald Knuth, and the other is the Foata normal form due to Pierre Cartier and Dominique Foata who studied the trace monoid for its combinatorics in the 1960s.[3]

Unicode's Normalization Form Canonical Decomposition (NFD) is an example of a lexicographic normal form - the ordering is to sort consecutive characters with non-zero canonical combining class by that class.

Trace languages

Just as a formal language can be regarded as a subset of Σ {\displaystyle \Sigma ^{*}} , the set of all possible strings, so a trace language is defined as a subset of M ( D ) {\displaystyle \mathbb {M} (D)} all possible traces.

Alternatively, but equivalently, a language L Σ {\displaystyle L\subseteq \Sigma ^{*}} is a trace language, or is said to be consistent with dependency D if

L = [ L ] D {\displaystyle L=[L]_{D}}

where

[ L ] D = w L [ w ] D {\displaystyle [L]_{D}=\bigcup _{w\in L}[w]_{D}}

is the trace closure of a set of strings.

See also

Notes

  1. ^ Sándor & Crstici (2004) p.161
  2. ^ Proposition 2.2, Diekert and Métivier 1997.
  3. ^ Section 2.3, Diekert and Métivier 1997.

References

General references

  • Diekert, Volker; Métivier, Yves (1997), "Partial Commutation and Traces", in Rozenberg, G.; Salomaa, A. (eds.), Handbook of Formal Languages Vol. 3; Beyond Words, Springer-Verlag, Berlin, pp. 457–534, ISBN 3-540-60649-1
  • Lothaire, M. (2011), Algebraic combinatorics on words, Encyclopedia of Mathematics and Its Applications, vol. 90, With preface by Jean Berstel and Dominique Perrin (Reprint of the 2002 hardback ed.), Cambridge University Press, ISBN 978-0-521-18071-9, Zbl 1221.68183
  • Antoni Mazurkiewicz, "Introduction to Trace Theory", pp 3–41, in The Book of Traces, V. Diekert, G. Rozenberg, eds. (1995) World Scientific, Singapore ISBN 981-02-2058-8
  • Volker Diekert, Combinatorics on traces, LNCS 454, Springer, 1990, ISBN 3-540-53031-2, pp. 9–29
  • Sándor, Jozsef; Crstici, Borislav (2004), Handbook of number theory II, Dordrecht: Kluwer Academic, pp. 32–36, ISBN 1-4020-2546-7, Zbl 1079.11001

Seminal publications

  • Pierre Cartier and Dominique Foata, Problèmes combinatoires de commutation et réarrangements, Lecture Notes in Mathematics 85, Springer-Verlag, Berlin, 1969, Free 2006 reprint with new appendixes
  • Antoni Mazurkiewicz, Concurrent program schemes and their interpretations, DAIMI Report PB 78, Aarhus University, 1977