Jump to content

Complexity function

From Wikipedia, the free encyclopedia
(Redirected from Morse–Hedlund theorem)

In computer science, the complexity function of a word or string (a finite or infinite sequence of symbols from some alphabet) is the function that counts the number of distinct factors (substrings of consecutive symbols) of that string. More generally, the complexity function of a formal language (a set of finite strings) counts the number of distinct words of given length.

Complexity function of a word

[edit]

Let u be a (possibly infinite) sequence of symbols from an alphabet. Define the function pu(n) of a positive integer n to be the number of different factors (consecutive substrings) of length n from the string u.[1][2][3][4][5]

For a string u of length at least n over an alphabet of size k we clearly have

the bounds being achieved by the constant word and a disjunctive word,[6] for example, the Champernowne word respectively.[7] For infinite words u, we have pu(n) bounded if u is ultimately periodic (a finite, possibly empty, sequence followed by a finite cycle). Conversely, if pu(n) ≤ n for some n, then u is ultimately periodic.[3][8]

An aperiodic sequence is one which is not ultimately periodic. An aperiodic sequence has strictly increasing complexity function (this is the Morse–Hedlund theorem),[9][10] so p(n) is at least n+1.[11]

A set S of finite binary words is balanced if for each n the subset Sn of words of length n has the property that the Hamming weight of the words in Sn takes at most two distinct values. A balanced sequence is one for which the set of factors is balanced.[12] A balanced sequence has complexity function at most n+1.[13]

A Sturmian word over a binary alphabet is one with complexity function n + 1.[14] A sequence is Sturmian if and only if it is balanced and aperiodic.[2][15] An example is the Fibonacci word.[14][16] More generally, a Sturmian word over an alphabet of size k is one with complexity n+k−1. An Arnoux-Rauzy word over a ternary alphabet has complexity 2n + 1:[14] an example is the Tribonacci word.[17]

For recurrent words, those in which each factor appears infinitely often, the complexity function almost characterises the set of factors: if s is a recurrent word with the same complexity function as t are then s has the same set of factors as t or δt where δ denotes the letter doubling morphism aaa.[18]

Complexity function of a language

[edit]

Let L be a language over an alphabet and define the function pL(n) of a positive integer n to be the number of different words of length n in L[9] The complexity function of a word is thus the complexity function of the language consisting of the factors of that word.

The complexity function of a language is less constrained than that of a word. For example, it may be bounded but not eventually constant: the complexity function of the regular language takes values 3 and 4 on odd and even n≥2 respectively. There is an analogue of the Morse–Hedlund theorem: if the complexity of L satisfies pL(n) ≤ n for some n, then pL is bounded and there is a finite language F such that[9]

A polynomial or sparse language is one for which the complexity function p(n) is bounded by a fixed power of n. A regular language which is not polynomial is exponential: there are infinitely many n for which p(n) is greater than kn for some fixed k > 1.[19]

[edit]

The topological entropy of an infinite sequence u is defined by

The limit exists as the logarithm of the complexity function is subadditive.[20][21] Every real number between 0 and 1 occurs as the topological entropy of some sequence is applicable,[22] which may be taken to be uniformly recurrent[23] or even uniquely ergodic.[24]

For x a real number and b an integer ≥ 2 then the complexity function of x in base b is the complexity function p(x,b,n) of the sequence of digits of x written in base b. If x is an irrational number then p(x,b,n) ≥ n+1; if x is rational then p(x,b,n) ≤ C for some constant C depending on x and b.[6] It is conjectured that for algebraic irrational x the complexity is bn (which would follow if all such numbers were normal) but all that is known in this case is that p grows faster than any linear function of n.[25]

The abelian complexity function pab(n) similarly counts the number of occurrences of distinct factors of given length n, where now we identify factors that differ only by a permutation of the positions. Clearly pab(n) ≤ p(n). The abelian complexity of a Sturmian sequence satisfies pab(n) = 2.[26]

References

[edit]
  1. ^ Lothaire (2011) p.7
  2. ^ a b Lothaire (2011) p.46
  3. ^ a b Pytheas Fogg (2002) p.3
  4. ^ Berstel et al (2009) p.82
  5. ^ Allouche & Shallit (2003) p.298
  6. ^ a b Bugeaud (2012) p.91
  7. ^ Cassaigne & Nicolas (2010) p.165
  8. ^ Allouche & Shallit (2003) p.302
  9. ^ a b c Berthé & Rigo (2010) p.166
  10. ^ Cassaigne & Nicolas (2010) p.166
  11. ^ Lothaire (2011) p.22
  12. ^ Allouche & Shallit (2003) p.313
  13. ^ Lothaire (2011) p.48
  14. ^ a b c Pytheas Fogg (2002) p.6
  15. ^ Allouche & Shallit (2003) p.318
  16. ^ de Luca, Aldo (1995). "A division property of the Fibonacci word". Information Processing Letters. 54 (6): 307–312. doi:10.1016/0020-0190(95)00067-M.
  17. ^ Pytheas Fogg (2002) p.368
  18. ^ Berstel et al (2009) p.84
  19. ^ Berthé & Rigo (2010) p.136
  20. ^ Pytheas Fogg (2002) p.4
  21. ^ Allouche & Shallit (2003) p.303
  22. ^ Cassaigne & Nicolas (2010) p.169
  23. ^ Berthé & Rigo (2010) p.391
  24. ^ Berthé & Rigo (2010) p.169
  25. ^ Berthé & Rigo (2010) p.414
  26. ^ Blanchet-Sadri, Francine; Fox, Nathan (2013). "On the Asymptotic Abelian Complexity of Morphic Words". In Béal, Marie-Pierre; Carton, Olivier (eds.). Developments in Language Theory. Proceedings, 17th International Conference, DLT 2013, Marne-la-Vallée, France, June 18-21, 2013. Lecture Notes in Computer Science. Vol. 7907. Berlin, Heidelberg: Springer-Verlag. pp. 94–105. doi:10.1007/978-3-642-38771-5_10. ISBN 978-3-642-38770-8. ISSN 0302-9743.