Combinatorics on words is a fairly new field of mathematics, branching from combinatorics, which focuses on the study of words and formal languages. The subject looks at letters or symbols, and the sequences they form. Combinatorics on words affects various areas of mathematical study, including algebra and computer science. There have been a wide range of contributions to the field. Some of the first work was on square-free words by Axel Thue in the early 1900s. He and colleagues observed patterns within words and tried to explain them. As time went on, combinatorics on words became useful in the study of algorithms and coding. It led to developments in abstract algebra and answering open questions.
Combinatorics is an area of discrete mathematics. Discrete mathematics is the study of countable structures. These objects have a definite beginning and end. The study of enumerable objects is the opposite of disciplines such as analysis, where calculus and infinite structures are studied. Combinatorics studies how to count these objects using various representations. Combinatorics on words is a recent development in this field that focuses on the study of words and formal languages. A formal language is any set of symbols and combinations of symbols that people use to communicate information.
Some terminology relevant to the study of words should first be explained. First and foremost, a word is basically a sequence of symbols, or letters, in a finite set.[1] One of these sets is known by the general public as the alphabet. For example, the word "encyclopedia" is a sequence of symbols in the English alphabet, a finite set of twenty-six letters. Since a word can be described as a sequence, other basic mathematical descriptions can be applied. The alphabet is a set, so as one would expect, the empty set is a subset. In other words, there exists a unique word of length zero. The length of the word
w
|w|
|w|=12
In addition to examining sequences in themselves, another area to consider of combinatorics on words is how they can be represented visually. In mathematics various structures are used to encode data. A common structure used in combinatorics is the tree structure. A tree structure is a graph where the vertices are connected by one line, called a path or edge. Trees may not contain cycles, and may or may not be complete. It is possible to encode a word, since a word is constructed by symbols, and encode the data by using a tree. This gives a visual representation of the object.
The first books on combinatorics on words that summarize the origins of the subject were written by a group of mathematicians that collectively went by the name of M. Lothaire. Their first book was published in 1983, when combinatorics on words became more widespread.
A main contributor to the development of combinatorics on words was Axel Thue (1863–1922); he researched repetition. Thue's main contribution was the proof of the existence of infinite square-free words. Square-free words do not have adjacent repeated factors. To clarify, "dining" is not square-free since "in" is repeated consecutively, while "servers" is square-free, its two "er" factors not being adjacent. Thue proves his conjecture on the existence of infinite square-free words by using substitutions. A substitution is a way to take a symbol and replace it with a word. He uses this technique to describe his other contribution, the Thue–Morse sequence, or Thue–Morse word.
Thue wrote two papers on square-free words, the second of which was on the Thue–Morse word. Marston Morse is included in the name because he discovered the same result as Thue did, yet they worked independently. Thue also proved the existence of an overlap-free word. An overlap-free word is when, for two symbols
x
y
As was previously described, words are studied by examining the sequences made by the symbols. Patterns are found, and they can be described mathematically. Patterns can be either avoidable patterns, or unavoidable. A significant contributor to the work of unavoidable patterns, or regularities, was Frank Ramsey in 1930. His important theorem states that for integers
k
m\geq2
Other contributors to the study of unavoidable patterns include van der Waerden. His theorem states that if the positive integers are partitioned into
k
c
c
When examining unavoidable patterns sesquipowers are also studied. For some patterns
x
y
z
x
\ldots~
Necklaces are constructed from words of circular sequences. They are most frequently used in music and astronomy. Flye Sainte-Marie in 1894 proved there are
2n-1-n | |
2 |
2n
In 1874, Baudot developed the code that would eventually take the place of Morse code by applying the theory of binary de Bruijn necklaces. The problem continued from Sainte-Marie to Martin in 1934, who began looking at algorithms to make words of the de Bruijn structure. It was then worked on by Posthumus in 1943.
See main article: Chomsky hierarchy. Possibly the most applied result in combinatorics on words is the Chomsky hierarchy, developed by Noam Chomsky. He studied formal language in the 1950s.[2] His way of looking at language simplified the subject. He disregards the actual meaning of the word, does not consider certain factors such as frequency and context, and applies patterns of short terms to all length terms. The basic idea of Chomsky's work is to divide language into four levels, or the language hierarchy. The four levels are: regular, context-free, context-sensitive, and computably enumerable or unrestricted. Regular is the least complex while computably enumerable is the most complex. While his work grew out of combinatorics on words, it drastically affected other disciplines, especially computer science.[3]
See main article: Sturmian word. Sturmian words, created by François Sturm, have roots in combinatorics on words. There exist several equivalent definitions of Sturmian words. For example, an infinite word is Sturmian if and only if it has
n+1
n
n
See main article: Lyndon word. A Lyndon word is a word over a given alphabet that is written in its simplest and most ordered form out of its respective conjugacy class. Lyndon words are important because for any given Lyndon word
x
y
z
y<z
x=yz
Cobham contributed work relating Eugène Prouhet's work with finite automata. A mathematical graph is made of edges and nodes. With finite automata, the edges are labeled with a letter in an alphabet. To use the graph, one starts at a node and travels along the edges to reach a final node. The path taken along the graph forms the word. It is a finite graph because there are a countable number of nodes and edges, and only one path connects two distinct nodes.
Gauss codes, created by Carl Friedrich Gauss in 1838, are developed from graphs. Specifically, a closed curve on a plane is needed. If the curve only crosses over itself a finite number of times, then one labels the intersections with a letter from the alphabet used. Traveling along the curve, the word is determined by recording each letter as an intersection is passed. Gauss noticed that the distance between when the same symbol shows up in a word is an even integer.
Walther Franz Anton von Dyck began the work of combinatorics on words in group theory by his published work in 1882 and 1883. He began by using words as group elements. Lagrange also contributed in 1771 with his work on permutation groups.
One aspect of combinatorics on words studied in group theory is reduced words. A group is constructed with words on some alphabet including generators and inverse elements, excluding factors that appear of the form aā or āa, for some a in the alphabet. Reduced words are formed when the factors aā, āa are used to cancel out elements until a unique word is reached. Nielsen transformations were also developed. For a set of elements of a free group, a Nielsen transformation is achieved by three transformations; replacing an element with its inverse, replacing an element with the product of itself and another element, and eliminating any element equal to 1. By applying these transformations Nielsen reduced sets are formed. A reduced set means no element can be multiplied by other elements to cancel out completely. There are also connections with Nielsen transformations with Sturmian words.
One problem considered in the study of combinatorics on words in group theory is the following: for two elements
x
y
x=y
x
y
The Burnside question was proved using the existence of an infinite cube-free word. This question asks if a group is finite if the group has a definite number of generators and meets the criteria
xn=1
x
g,h
w
g(w)=h(w)
Combinatorics on words have applications on equations. Makanin proved that it is possible to find a solution for a finite system of equations, when the equations are constructed from words.