# Organization:Association scheme

The theory of association schemes arose in statistics, in the theory of experimental design for the analysis of variance. In mathematics, association schemes belong to both algebra and combinatorics. In algebraic combinatorics, association schemes provide a unified approach to many topics, for example combinatorial designs and coding theory. In algebra, association schemes generalize groups, and the theory of association schemes generalizes the character theory of linear representations of groups.

## Definition

An n-class association scheme consists of a set X together with a partition S of X × X into n + 1 binary relations, R0, R1, ..., Rn which satisfy:

• $\displaystyle{ R_{0} = \{(x,x) : x \in X\} }$ and is called the identity relation.
• Defining $\displaystyle{ R^* := \{(x,y) : (y,x) \in R\} }$, if R in S, then R* in S
• If $\displaystyle{ (x,y) \in R_{k} }$, the number of $\displaystyle{ z \in X }$ such that $\displaystyle{ (x,z) \in R_{i} }$ and $\displaystyle{ (z,y) \in R_{j} }$ is a constant $\displaystyle{ p^k_{ij} }$ depending on $\displaystyle{ i }$, $\displaystyle{ j }$, $\displaystyle{ k }$ but not on the particular choice of $\displaystyle{ x }$ and $\displaystyle{ y }$.

An association scheme is commutative if $\displaystyle{ p_{ij}^k = p_{ji}^k }$ for all $\displaystyle{ i }$, $\displaystyle{ j }$ and $\displaystyle{ k }$. Most authors assume this property.

A symmetric association scheme is one in which each $\displaystyle{ R_i }$ is a symmetric relation. That is:

• if (x, y) ∈ Ri, then (y, x) ∈ Ri. (Or equivalently, R* = R.)

Every symmetric association scheme is commutative.

Note, however, that while the notion of an association scheme generalizes the notion of a group, the notion of a commutative association scheme only generalizes the notion of a commutative group.

Two points x and y are called i th associates if $\displaystyle{ (x,y) \in R_i }$. The definition states that if x and y are i th associates then so are y and x. Every pair of points are i th associates for exactly one $\displaystyle{ i }$. Each point is its own zeroth associate while distinct points are never zeroth associates. If x and y are k th associates then the number of points $\displaystyle{ z }$ which are both i th associates of $\displaystyle{ x }$ and j th associates of $\displaystyle{ y }$ is a constant $\displaystyle{ p^k_{ij} }$.

### Graph interpretation and adjacency matrices

A symmetric association scheme can be visualized as a complete graph with labeled edges. The graph has $\displaystyle{ v }$ vertices, one for each point of $\displaystyle{ X }$, and the edge joining vertices $\displaystyle{ x }$ and $\displaystyle{ y }$ is labeled $\displaystyle{ i }$ if $\displaystyle{ x }$ and $\displaystyle{ y }$ are $\displaystyle{ i }$ th associates. Each edge has a unique label, and the number of triangles with a fixed base labeled $\displaystyle{ k }$ having the other edges labeled $\displaystyle{ i }$ and $\displaystyle{ j }$ is a constant $\displaystyle{ p^k_{ij} }$, depending on $\displaystyle{ i,j,k }$ but not on the choice of the base. In particular, each vertex is incident with exactly $\displaystyle{ p^0_{ii}=v_{i} }$ edges labeled $\displaystyle{ i }$; $\displaystyle{ v_{i} }$ is the valency of the relation $\displaystyle{ R_{i} }$. There are also loops labeled $\displaystyle{ 0 }$ at each vertex $\displaystyle{ x }$, corresponding to $\displaystyle{ R_{0} }$.

The relations are described by their adjacency matrices. $\displaystyle{ A_i }$ is the adjacency matrix of $\displaystyle{ R_{i} }$ for $\displaystyle{ i=0,\ldots,n }$ and is a v × v matrix with rows and columns labeled by the points of $\displaystyle{ X }$.

$\displaystyle{ \left( A_i \right)_{x,y} = \begin{cases} 1, & \mbox{if } (x,y) \in R_{i},\\ 0, & \mbox{otherwise.} \end{cases}\qquad (1) }$

The definition of a symmetric association scheme is equivalent to saying that the $\displaystyle{ A_i }$ are v × v (0,1)-matrices which satisfy

I. $\displaystyle{ A_i }$ is symmetric,
II. $\displaystyle{ \sum_{i=0}^n A_i = J }$ (the all-ones matrix),
III. $\displaystyle{ A_0 = I }$,
IV. $\displaystyle{ A_i A_j = \sum_{k=0}^n p^k_{ij}A_k = A_j A_i, i,j=0,\ldots,n }$.

The (x, y)-th entry of the left side of (IV) is the number of paths of length two between x and y with labels i and j in the graph. Note that the rows and columns of $\displaystyle{ A_{i} }$ contain $\displaystyle{ v_{i} }$ $\displaystyle{ 1 }$'s:

$\displaystyle{ A_{i} J=J A_{i}=v_{i} J. \qquad(2) }$

### Terminology

• The numbers $\displaystyle{ p_{ij}^k }$ are called the parameters of the scheme. They are also referred to as the structural constants.

## History

The term association scheme is due to (Bose Shimamoto) but the concept is already inherent in (Bose Nair). These authors were studying what statisticians have called partially balanced incomplete block designs (PBIBDs). The subject became an object of algebraic interest with the publication of (Bose Mesner) and the introduction of the Bose–Mesner algebra. The most important contribution to the theory was the thesis of P. Delsarte (Delsarte 1973) who recognized and fully used the connections with coding theory and design theory. Generalizations have been studied by D. G. Higman (coherent configurations) and B. Weisfeiler (distance regular graphs).

## Basic facts

• $\displaystyle{ p_{00}^0 = 1 }$, i.e. if $\displaystyle{ (x,y) \in R_0 }$ then $\displaystyle{ x = y }$ and the only $\displaystyle{ z }$ such that $\displaystyle{ (x,z) \in R_0 }$ is $\displaystyle{ z=x }$.
• $\displaystyle{ \sum_{i=0}^{k} p_{ii}^0 = |X| }$, this is because the $\displaystyle{ R_i }$ partition $\displaystyle{ X }$.

## The Bose–Mesner algebra

The adjacency matrices $\displaystyle{ A_i }$ of the graphs $\displaystyle{ \left(X,R_{i}\right) }$ generate a commutative and associative algebra $\displaystyle{ \mathcal{A} }$ (over the real or complex numbers) both for the matrix product and the pointwise product. This associative, commutative algebra is called the Bose–Mesner algebra of the association scheme.

Since the matrices in $\displaystyle{ \mathcal{A} }$ are symmetric and commute with each other, they can be diagonalized simultaneously. Therefore, $\displaystyle{ \mathcal{A} }$ is semi-simple and has a unique basis of primitive idempotents $\displaystyle{ J_{0},\ldots,J_{n} }$.

There is another algebra of $\displaystyle{ (n+1)\times(n+1) }$ matrices which is isomorphic to $\displaystyle{ \mathcal{A} }$, and is often easier to work with.

## Examples

• The Johnson scheme, denoted J(v, k), is defined as follows. Let S be a set with v elements. The points of the scheme J(v, k) are the $\displaystyle{ {v \choose k} }$ subsets of S with k elements. Two k-element subsets A, B of S are i th associates when their intersection has size k − i.
• The Hamming scheme, denoted H(n, q), is defined as follows. The points of H(n, q) are the qn ordered n-tuples over a set of size q. Two n-tuples x, y are said to be i th associates if they disagree in exactly i coordinates. E.g., if x = (1,0,1,1), y = (1,1,1,1), z = (0,0,1,1), then x and y are 1st associates, x and z are 1st associates and y and z are 2nd associates in H(4,2).
• A distance-regular graph, G, forms an association scheme by defining two vertices to be i th associates if their distance is i.
• A finite group G yields an association scheme on $\displaystyle{ X=G }$, with a class Rg for each group element, as follows: for each $\displaystyle{ g \in G }$ let $\displaystyle{ R_g = \{(x,y) \mid x=g*y\} }$ where $\displaystyle{ * }$ is the group operation. The class of the group identity is R0. This association scheme is commutative if and only if G is abelian.
• A specific 3-class association scheme:
Let A(3) be the following association scheme with three associate classes on the set X = {1,2,3,4,5,6}. The (i, j ) entry is s if elements i and j are in relation Rs.
1 2 3 4 5 6
1  0   1   1   2   3   3
2  1   0   1   3   2   3
3  1   1   0   3   3   2
4  2   3   3   0   1   1
5  3   2   3   1   0   1
6  3   3   2   1   1   0

## Coding theory

The Hamming scheme and the Johnson scheme are of major significance in classical coding theory.

In coding theory, association scheme theory is mainly concerned with the distance of a code. The linear programming method produces upper bounds for the size of a code with given minimum distance, and lower bounds for the size of a design with a given strength. The most specific results are obtained in the case where the underlying association scheme satisfies certain polynomial properties; this leads one into the realm of orthogonal polynomials. In particular, some universal bounds are derived for codes and designs in polynomial-type association schemes.

In classical coding theory, dealing with codes in a Hamming scheme, the MacWilliams transform involves a family of orthogonal polynomials known as the Krawtchouk polynomials. These polynomials give the eigenvalues of the distance relation matrices of the Hamming scheme.