Jump to ContentJump to Main Navigation
Show Summary Details
More options …

Open Mathematics

formerly Central European Journal of Mathematics

Editor-in-Chief: Gianazza, Ugo / Vespri, Vincenzo


IMPACT FACTOR 2018: 0.726
5-year IMPACT FACTOR: 0.869

CiteScore 2018: 0.90

SCImago Journal Rank (SJR) 2018: 0.323
Source Normalized Impact per Paper (SNIP) 2018: 0.821

Mathematical Citation Quotient (MCQ) 2018: 0.34

ICV 2017: 161.82

Open Access
Online
ISSN
2391-5455
See all formats and pricing
More options …
Volume 16, Issue 1

Issues

Volume 13 (2015)

Spectrum of free-form Sudoku graphs

Mohammad Abudayah / Omar Alomari / Torsten Sander
  • Corresponding author
  • Ostfalia Hochschule für angewandte Wissenschaften, Fakultät für Informatik, Wolfenbüttel, Germany
  • Email
  • Other articles by this author:
  • De Gruyter OnlineGoogle Scholar
Published Online: 2018-12-31 | DOI: https://doi.org/10.1515/math-2018-0125

Abstract

A free-form Sudoku puzzle is a square arrangement of m ×m cells such that the cells are partitioned into m subsets (called blocks) of equal cardinality. The goal of the puzzle is to place integers 1, . . , m in the cells such that the numbers in every row, column and block are distinct. Represent each cell by a vertex and add edges between two vertices exactly when the corresponding cells, according to the rules, must contain different numbers. This yields the associated free-form Sudoku graph. This article studies the eigenvalues of free-form Sudoku graphs, most notably integrality. Further, we analyze the evolution of eigenvalues and eigenspaces of such graphs when the associated puzzle is subjected to a ‘blow up’ operation, which scales the cell grid including its block partition.

Keywords: Sudoku; Spectrum; Eigenvectors

MSC 2010: 05C50; 15A18

1 Introduction

The recreational game of Sudoku has been popular for several years now. Its classic variant is played on a board with 9 × 9 cells, subdivided into a 3 × 3 grid of square blocks containing 3 × 3 cells each. Each cell may be empty or contain one of the numbers 1, . . , 9. A number of cells of each puzzle will have been pre-filled by the puzzle creator. The goal of the puzzle solver is to fill the remaining cells with the numbers 1, . . , 9 such that in the completed puzzle the number in each cell occurs only once per row, column and block.

Let us call the classical variant the 3-Sudoku. Its 32×32 board contains 34 cells. One can readily generalize the game to n-Sudokus played on an n2 × n2 board with n4 cells, subdivided into n2 square blocks with n2 cells in each. The permitted numbers in the cells now range from 1, . . , n2, but the restrictions for a valid solution remain the same. The 2-Sudoku is also known as a Shidoku.

Despite the seemingly recreational character of the game, it offers a surprising number of mathematical facets. This makes Sudoku an interesting topic for mathematics lessons [1, 2]. Among the topics touched by Sudoku are problem solving, latin squares, counting, exhausting symmetry and colouring problems on graphs (see, for example, the introductory book [3]). Due to this, more and more interesting results about Sudoku have been published in the recent past. For instance, different approaches for solvers have been presented in [4, 5, 6]. The combinatorial properties of completed Sudoku squares as a family of Latin squares, in particular the search for orthogonal pairs, have been considered in [7, 8, 9, 10]. Other researchers focus on algebraic aspects of Sudoku, especially groups and rings associated with Sudoku (cf. [11, 12, 13]).

One of the most intuitive links between Sudoku and mathematics is that the processes of solving a Sudoku can be interpreted as completing a given partial vertex colouring of a certain graph (each preassigned colour corresponds to a prefilled number in a cell, a so-called clue). To this end, we represent each cell of the given Sudoku square by a single vertex. Two vertices are adjacent if and only if the associated cells must not contain the same number (according to the rules of the game). Figure 1 (a) depicts the adjacencies in the neighborhood of an exemplary vertex of the 2-Sudoku graph. Using this representation, the valid solutions of a given n-Sudoku can be characterized as proper optimal vertex colourings (using n2 colors each) of the associated Sudoku graph.

Deriving the graph of a classic and a free-form 2-Sudoku puzzle
Figure 1

Deriving the graph of a classic and a free-form 2-Sudoku puzzle

Sudoku graphs have also been studied from the perspective of algebraic graph theory. It has been shown that every Sudoku graph is integral, i.e., all eigenvalues of its adjacency matrix are integers. This has been shown algebraically by means of group characters [14], but it also follows from the fact that Sudoku graphs are actually NEPS of complete graphs. Hence they belong to the class of gcd-graphs, a subclass of the integral Cayley graphs over abelian groups [15]. Here, NEPS is the common short form of the non-complete extended p-sum, a generalized graph product that includes many known products [16].

There exist many generalizations and variants of the classic Sudoku, the n-Sudokus being the most common one. Besides changing the size of a Sudoku one can also try to vary every other aspect of a Sudoku, e.g. change the rules, introduce additional rules or change the shape of the blocks. The book [3] presents many such variants. Changing the shape of the blocks leads to the notion of a free-form Sudoku. Given a square arrangement of m ×m cells (where m need not be a square number any more), we permit the blocks to be an arbitrary partition T of the cells into m subsets of equal cardinality. Note that the blocks are not required to be contiguous arrangements of cells. The Sudoku rules remain unchanged, in the sense that we need to fill in the numbers 1, . . , m such that each row, column and block contains m distinct numbers. We shall denote the associated graph by FSud(m, T). Figure 1 (b) illustrates the adjacencies in the neighborhood of an exemplary vertex for the graph of a free-form 2-Sudoku. The cells of the (non-contiguous) block containing the considered vertex have been shaded.

It seems that free-form Sudokus have not been studied in the literature so far although they are included in many Sudoku puzzle books that offer challenging variants. Computer experiments readily indicate that integrality of the eigenvalues critically depends on the degree of symmetry exhibited by the chosen cell partition. However, free-form Sudoku graphs are well structured enough to predict the changes of their spectra and eigenspaces exactly, when they are transformed in certain ways.

2 Integrality

In the following, consider a given free-form Sudoku graph FSud(m, T) and let A be its adjacency matrix (with respect to some arbitrary but fixed vertex order). The structure of the adjacency matrix of a free-form Sudoku graph is governed by the rules of Sudoku:

Observation 2.1

Two vertices of a free-form Sudoku graph are adjacent if and only if one of the following mutually exclusive places applies:

  • (B)

    The associated cells belong to the same block of the tiling

  • (H)

    The associated cells belong to the same row of the puzzle, but not to the same block

  • (V)

    The associated cells belong to the same column of the puzzle, but not to the same block

The chosen partition into mutually exclusive cases immediately gives rise to a decomposition of the adjacency matrix according to these three cases:

A=LB+LH+LVRm2×m2(1)

In this sense, the graph FSud(m, T) can be interpreted as a composition of three layers of edges, according to the cases (B), (H), (V). These layers can be viewed and studied as graphs in their own right.

Proposition 2.2

For i, j ϵ {1, . . , m} let pij denote the number of cells in the i-th row (resp. column) of the puzzle that belong to the j-th block of the partition. Then L H (resp. LV) is an adjacency matrix of the union of m disjoint complete multipartite graphs Kpi1,..,pim , i = 1, . . , m

Proof. Consider a fixed row i of the given Sudoku puzzle.We group the cells of the considered row according to their respective block membership. In view of rule (H)we see that the vertices corresponding to the grouped cells are adjacent if and only if they belong to different blocks. So the groups of cells form m independent sets of the respective sizes pi1, . . , pim, further all edges exist between different groups. The reasoning for LV is analogous; just consider a fixed column and rule (V).

Proposition 2.3

L B is an adjacency matrix of the union of m disjoint complete graphs Km

Proof. Group the vertices according to block membership of their associated cells. Since each block contains m cells the result follows from rule (B).

Proposition 2.4

The following statements are equivalent:

  • (a)

    LH (resp. LV) represents a regular graph

  • (b)

    LH (resp. LV) has constant row sum

  • (c)

    LB commutes with LH (resp. LV)

Proof. Statements (a) and (b) are clearly equivalent. Next we show that statement (b) implies (c). To this end, note that (LH)is(LB)sj = 1 if and only if (LH)is = (LB)sj = 1. This translates to the requirement that the cells associated with vertices i and s are in the same row (but not the same block) and that the cells associated with vertices s and j belong to the same block. Therefore, (LHLB)ij counts the vertices s that meet the mentioned requirement for given i, j. In a similar manner we see that (LBLH)ij counts the vertices s such that the cells associated with vertices i and s belong to same block and that the cells associated with vertices s and j are in the same row (but do not belong to the same block). If, however, L H has constant row sum, then it follows by rule (H) that each cell has the same common number of cells that are in the same row but not in the same block as the considered cell. Equivalently, since all rows have the same number of cells, each cell has the same common number k of cells that are in the same row and in the same block as the considered cell. Therefore, any block that has cells in a row must have exactly k cells in that row. Hence the two counts mentioned before are identical: (LHLB)ij = (LBLH)ij.

Conversely, assume that statement (c) is true. Consider two vertices i and j such that their associated cells belong to different rows and different blocks. Then (LBLH)ij = (LHLB)ij implies that these two blocks have the same number of cells in the two rows. Hence there exists a common number k such that any block that has cells in a row must have exactly k cells in that row. Considering a single row of L H, it follows that it has row sum mk since the puzzle row of the cell associated with that matrix row contains m cells and among them k cells that belong to the same block as the considered cell.

Considering the third statement of Proposition 2.4, note that L H and L V need not commute even when both LH and LV have constant row sum, cf. the graph FSud(4, T) for T = {{1, 8, 9, 16}, {3, 6, 10, 15}, {2, 7, 11, 14}, {4, 5, 12, 13}} (here we number the cells from left to right, row after row).

Theorem 2.5

(see [17]). The characteristic polynomial of the complete multipartite graph G = Kp1,..,pk is

χ(G)=xnkxkm=2k(m1)σmxkm,

where σ1=1ikpi,σ2=1i<jkpipj,σ3=1i<j<lkpipjpl and so on up to σk=1ikpi.

Corollary 2.6

Let G = Kq,..,q be a complete k-partite graph. Then its spectrum is

σ(G)=0(kqk),(k1)q(1),q(k1).

Proof.

χ(G)=xkqkxkk2q2xk22q3k3xk33q4k4xk4(k1)qk=xkqkx(k1)qxk1+k11xk2q+k12xk3q2++qk1=xkqk(x(k1)q)(x+q)k1.

Theorem 2.7

Given a free-form Sudoku puzzle, assume that the following conditions are met:

  • (i)

    There exists a common number q such that for each cell there exist exactly q cells belonging to the same row and block as the considered cell (including itself)

  • (ii)

    A similar condition holds with ‘row’ replaced by ‘column’

  • (iii)

    For any two cells C1, C2 belonging to different rows and columns, the unique cell lying in the same row as C1 and the same column as C2 belongs to the same block as C1 if and only if the unique cell lying in the same row as C2 and the same column as C1 belongs to the same block as C2.

Then the associated graph FSud(m, T) is integral

Proof. Under condition (i) we have pij ϵ {0, q} for all i, j ϵ {1, . . , m} in Proposition 2.2. Therefore, LH represents a union of complete multipartite graphs Kq,..,q. Hence L H is integral by Corollary 2.6. Likewise, condition (ii) ensures that LV is integral. L B is integral by Proposition 2.3 and Corollary 2.6 (for r = 1).

We can conclude from Proposition 2.4 and conditions (i) and (ii) that the pairs L H, L B and LV , L B commute. Condition (iii) is equivalent to the condition that L H and LV commute. Hence it follows that L H, LV, L B are simultaneously diagonizable. Thus their sum A = L H + LV + L B is integral as well.

It is easily checked that the classical n-Sudokus fulfill the conditions of Theorem 2.7. Hence the theorem provides a new way of proving integrality of classical Sudoku graphs (e.g. different from the proofs in [14], [15]). Further, Theorem 2.7 helps us to identify many more integral free-form Sudoku graphs besides the classical variants.

3 Transformations

In this section we consider a special kind of transformation of Sudoku puzzles and study its spectral properties. To this end, we define the k-fold blow up of a free-form Sudoku. This is formed by replacing each cell of the original Sudoku by a k × k arrangement of cells. The block partition Tʹ of the new graph is derived from the original partition T as follows. For each block B ϵ T we create a block Bʹ ϵ Tʹ by collecting the replacement cells of all the cells in B. In terms of graphs we see that the k-fold blow up transforms FSud(n, T) into FSud(kn, Tʹ). Let us denote the latter graph by FSud↑k(n, T). In Figure 2 the 3-fold blow up of a free-form 2-Sudoku is illustrated. It also sketches the neighborhood of an exemplary vertex of the blown up graph.

3-fold blow up of the free-form 2-Sudoku puzzle from Figure 1 (b)
Figure 2

3-fold blow up of the free-form 2-Sudoku puzzle from Figure 1 (b)

Let us now investigate how the adjacency matrix of a blown up Sudoku graph can be determined from the adjacency matrix of the original graph. To this end, we a assume that a fixed but otherwise arbitrary numbering of the cells of the original free-form Sudoku is given. The vertices of the associated graph shall be numbered accordingly. Further, we number the cells of a k-fold blow up of the given puzzle as follows. Let Si denote the k × k subsquare of the blown up puzzle containing exactly the replacement cells of the original cell number i.We number the k2n2 cells by subsequently numbering all vertices in S1, then S2 and so on. Inside each Si we number the cells by starting in the top left corner and proceeding from left to right, advancing row by row.

Next, note the following facts about the blow up operation:

Observation 3.1

  • If two cells i and j lie in the same row (resp. column) of the original puzzle, then all cells sharing the same relative row (resp. column) index inside Si and/or Sj lie in the same row (resp. column) of the blown up puzzle

  • If two cells i and j lie in the same block of the original puzzle, then all cells from Si and Sj lie in the same block of the blown up puzzle

In view of these facts and due to the chosen cell numbering of the blown up puzzle we can construct the adjacency matrix A↑ of FSud↑k(n, T) from the adjacency matrix A = (aij) of FSud(n, T) by replacing each entry aij of A by a (k2 × k2)-matrix that solely depends on whether (in the original puzzle) cell number j is in the same block as cell number i, or otherwise in the same row or column as j or lies somewhere else.

For this purpose we define the symbolic template adjacency matrix T(A) = (tij) as follows:

  • tii = D,

  • tij = B if i/ = j and cells i, j are in the same block of the original puzzle,

  • tij = H if cells i, j are not in the same block but in the same row of the original puzzle,

  • tij = V if cells i, j are not in the same block but in the same column of the original puzzle,

  • tij = N else.

The next step is to define the matrices that will be substituted into the template matrix. But first we need some building blocks. Let Ir denote the identity matrix, Jr the all-ones matrix and Nr the zero matrix of size r × r. Further, we will make use of the Kronecker product of real matrices, cf. [18].

Fixing the given blow up factor k, we define the following (k2 × k2)-matrices:

H=IkJk,V=JkIk,B=Jk2,D=Jk2Ik2,N=Nk2.(2)

The following result should now be self-evident:

Proposition 3.2

Given an n × n free-form Sudoku puzzle with block partition T, the adjacency matrix of FSud↑k(n, T) (with respect to the vertex numbering mentioned in the second paragraph of this section) can be obtained from the symbolic template matrix T(A) = (tij) of the adjacency matrix A = (aij) of the graph FSud(n, T) by replacing each entry tij of T(A) by the contents of the matching matrix from (2) that has the same name as the symbolic value of tij suggests

The replacement process can be expressed by means of the Kronecker product. Let us take the template matrix T(A) and use it to partition the non-zero entries aij of A according to their respective symbols tij. Setting the blown up adjacency matrix A↑ can now be expressed as follows:

Proposition 3.3

A=LBB+LHH+LVV+LDD.(3)

Our goal is to study the eigenvalues of blown up free-form Sudokus and express them in terms of the eigenvalues of the original puzzle.

For the rest of this section, we consider an arbitrary but fixed free-form n × n Sudoku puzzle with tiling T. Let A be the adjacency matrix of its associated graph FSud(n, T) and let A↑ be the adjacency matrix of FSud↑k(n, T).We assume that LV, L H, L B etc. denote the matrices appearing in equations (1) and (3). Further, let Eig(M) represent the set of all eigenvectors of a given matrix (or even a graph) M and let ker(M) be the set of all eigenvectors of M associated with eigenvalue 0. Note that neither set contains the null vector.

In the following, we will make use of the following properties of the Kronecker product:

Theorem 3.4

(see [18]).

  1. (αA) ⊗ B = A ⊗ (αB) = α(AB) for all α ϵ ℝ, A ϵ ℝp×q , B ϵ ℝr×s

  2. (AB) ⊗ C = A ⊗ (BC) for all A ϵ ℝm×n , B ϵ ℝp×q , C ϵ ℝr×s

  3. (A + B) ⊗ C = AC + BC for all A, B ϵ ℝp×q , C ϵ ℝr×s

  4. A ⊗ (B + C) = AB + AC for all A ϵ ℝp×q , B, C ϵ ℝr×s

  5. (AB)(CD) = ACBD for all A ϵ ℝp×q , B ϵ ℝr×s , C ϵ ℝq×k , D ϵ ℝs×l

Note that the Kronecker product formally includes the case where one or both factors are vectors.

Lemma 3.5

Suppose that x and y are eigenvectors of LV and Jk corresponding to the eigenvalues λ and α, respectively. Then, for any z ϵ ker(Jk), xyz is an eigenvector of A corresponding to eigenvalue λα − 1.

Proof. We use equations (2), (3) and the facts LV x = λx, Jky = αy, Jkz = 0.

A(xyz)=(LBB+LHH+LVV+LDD)(xyz)=(LBJK2+LHIkJk+LVJkIk+LDJk2LDIk2)(xyz)=((LBx)(Jky)(Jkz))+((LHx)(Iky)(Jkz))+((LVx)(Jky)(Ikz))+((LDx)(Jky)(Jkz))((LDx)(Iky)(Ikz))=(λx)(αy)zxyz=(λα1)(xyz).

Lemma 3.6

Suppose that x and z are eigenvectors of LH and Jk corresponding to the eigenvalues λ and α, respectively. Then, for any y ϵ ker(Jk), xyz is an eigenvector of A↑ corresponding to eigenvalue λα − 1.

Proof. Similar to the proof of Lemma 3.5.

The previous two lemmas will play a role in the construction of a basis of eigenvectors of a blown up Sudoku graph. To this end, we need to know the intersection of the spans of the two vector sets mentioned there. In the following, let 1r denote the all-ones vector of dimension r. Now note the following obvious facts:

Proposition 3.7

1. The spectrum of Jk is {k(1), 0(k−1)}.

2. The set

KJ:={(1,1,0,0,...,0,0)T,(1,0,1,0,...,0,0)T,...,(1,0,0,0,...,0,1)T}

is a maximal linearly independent subset of ker(Jk).

3. The set {1k} ∪ 𝒦J is a maximal linearly independent subset of Eig(Jk).

4. 1k ker(Jn).

For the next lemma we define the following sets:

XH={xyz:xEig(LH),yKer(Jk),zEig(Jk)},XV={xyz:xEig(LV),yEig(Jk),zKer(Jk)},X~H={xyz:xEig(LH),y,zKer(Jk)},X~V={xyz:xEig(LV),y,zKer(Jk)}.

Lemma 3.8

span(XH)span(XV)=span(X~H)=span(X~V).

Proof. Since ker(Jk) ⊆ Eig(Jk) it is obvious that

span(X~H)span(XH)span(XV).

Conversely, note that both LH and LV are symmetric and therefore diagonizable, i.e. span(Eig(LH)) = span(Eig(LH))=span(Eig(LV))=Rn2. Using this we conclude

span(XH)Rn2span(Ker(Jk))span(Eig(Jk)),span(XV)Rn2span(Eig(Jk))span(Ker(Jk)),span(X~H)=Rn2span(Ker(Jk))span(Ker(Jk)).

Once again employing the fact that ker(Jk) ⊆ Eig(Jk), we arrive at

span(XH)span(XV)span(X~H).

Lemma 3.9

Let x ϵ ℝn2 and y, z ϵ ker(Jk). Then xyz is an eigenvector of A corresponding to eigenvalue −1.

Proof.

A(xyz)=(LBJK2+LHIkJk+LVJkIk+LDJk2LDIk2)(xyz)=((LBx)(Jky)(Jkz))+((LHx)(Iky)(Jkz))+((LVx)(Jky)(Ikz))+((LDx)(Jky)(Jkz))((LDx)(Iky)(Ikz))=((In2x)(Iky)(Ikz))=(xyz).

Lemma 3.10

For any eigenvector x of the matrix k2LB + kLH + kLV corresponding to eigenvalue λ, the vector x1k2is an eigenvector of A corresponding to eigenvalue λ + k2 − 1.

Proof.

Ax1k2=LBB+LHH+LVV+LDDx1k2=LBxJk21k2+LHxH1k2+LVxV1k2+LDxJk21k2LDxJk21k2=LBxk21k2+LHxk1k2+LVxk1k2+LDxk21k2LDx1k2k2LBx1k2+kLHx1k2+kLVx1k2+k21LDx1k2=k2LB+kLv+kLvx1k2+k21x1k2=λ+k21x1k2.

Before we present the main result of this section we need one more technical lemma.

Lemma 3.11

Let {Yi}i=1nbe a set of linearly independent vectors. Then, for any set {Xi}i=1mof nonzero vectors and any function

ϕ:{1,2,3,,n}{1,2,3,,m},

the set {Xϕ(i)Yi}i=1nis linearly independent

Proof. Suppose otherwise that

i=1nci(Xϕ(i)Yi)=0

for suitable numbers ci ϵ ℝ. Let X ϕ(i) = (x1,ϕ(i), . . , xr,ϕ(i))T. By the definition of the Kronecker product we have

i=1nci(Xϕ(i)Yi)=(i=1ncix1,ϕ(i)Yii=1ncixr,ϕ(i)Yi)=0.

Thus, for each j we have

i=1ncixj,ϕ(i)Yi=0.

Due to the linear independence of the vectors Yi we see that cixj,ϕ(i) = 0 for every j = 1, . . , r and i = 1, . . , n. Now assume that ci*/ = 0 for some index i*. Then xj,= 0 for all j, therefore X = 0. But this is impossible {Xi}i=1mϕ(i*) ϕ(i*) since is a set of nonzero vectors.

For what follows, let 𝓑V, 𝓑H and 𝓑M denote arbitrary maximal linearly independent subsets of Eig(LV)2 , Eig(LH) and Eig(k2LB + kLV + kLH), respectively. Further, let ℰ = {e1, . . , en2} be the standard basis of ℝn, where ei denotes the i-th unit vector.

Theorem 3.12

Given a graph FSud(n, T), define the sets

χV={x1ky:xBV,yKJ},χH={xy1k:xBH,yKJ},χE={xyz:xε,y,zKJ},χM={x1k2:xBM}.

Then, their union 𝒳V ∪ 𝒳H ∪ 𝒳E ∪ 𝒳M forms a maximal linearly independent subset of Eig(FSud↑k(n, T)).

Proof. By construction and Lemma 3.1., each of the sets 𝒳V, 𝒳H, 𝒳E, 𝒳M in itself is linearly independent. Further, by construction and Proposition 3.7, the spans of these four sets are mutually disjoint (neglecting the null vector). Moreover, Lemmas 3.5, 3.6, 3.9 and 3.11 guarantee that the union contains only eigenvectors of FSud↑k(n, T). Finally, note that

|BV|=|BH|=|ε|=n2,|Kj|=k1

so that

|χV|=|χH|=(k1)n2,|χE|=|χH|=(k1)n2,|χM|=n2

and thus

|χV|+|χH|+|χE|+|χM|=2(k1)n2+(k1)2n2+n2=k2n2

Looking closer at Theorem 3.12, we see that if LV, L H, k2LB + kL H + kLV were all integral, then FSud↑k(n, T) would be integral as well.

Corollary 3.13

Under the conditions stated in Theorem 2.7 it follows that the blown up graph FSud↑k(n, T) is integral for every k

Proof. From the proof of Theorem 2.7 it follows that LV, L H and L B are all integral and commute with each other, hence they are simultaneously diagonizable. Consequently, k2LB + kL H + kLV is integral and therefore so is FSud↑k(n, T).

Owing to Theorem 3.12, we can use Lemmas 3.5, 3.6, 3.9 and 3.11 to establish the spectrum of a k-fold blow up from its original. Interestingly, we can predict explicitly that the largest eigenvalue stems from the set 𝒳M:

Theorem 3.14

Given a graph FSud(n, T), let λ be the largest eigenvalue of the associated matrix k2LB + kLV + kL H. Then λ + k2 − 1 is the largest eigenvalue of FSud↑k(n, T).

Proof. For the purposes of this proof we renumber the vertices of FSud(n, T) such that we sequentially number the vertices with one block, then continue with the next block and so on. The order in which the vertices are numbered with a single block is arbitrary. With respect to this vertex order the matrix L B assumes the form In JnIn2 . Clearly, this matrix contains JnIn as a principal submatrix. Consequently, the matrix M := k2LB + kLV + kL H contains the matrix k2Jnk2In as a principal submatrix, the latter having maximum eigenvalue k2(n − 1). By virtue of eigenvalue interlacing (see e.g. [19]) we conclude that λ > (n − 1)k2. So, according to Lemma 3.10, the largest eigenvalue of FSud↑k(n, T) originating from the set 𝒳M is at least (n − 1)k2 + k2 − 1 = nk2 − 1. We will now show that the largest eigenvalues originating from 𝒳V, 𝒳H and 𝒳B are smaller than this bound.

Since the largest eigenvalue of a matrix is bounded from above by the maximum row sum of the matrix, it is clear that the maximum eigenvalue both of L H and LV can be at most n − 1. Now recall from Proposition 3.7 that k is the maximum eigenvalue of Jk. Combining these findings, it now follows from Lemmas 3.5 and 3.6 that none of the eigenvalues associated with the vectors of the sets 𝒳V and 𝒳H exceeds (n − 1)k − 1, which is less than the lower bound given for 𝒳M. Finally, Lemma 3.9 tells us that no positive eigenvalue of FSud↑k(n, T) originates from 𝒳B.

4 Conclusion

Up to now, it seems that free-form Sudokus have not been researched at all. Providing a starting point, we have studied integrality of these graphs. Moreover, we have presented the blow up operation and shown how to obtain the eigenvalues of blown up free-form Sudokus from their originals. We would like to inspire more research on this topic, in particular regarding further spectral properties of free-form Sudoku graphs. Let us therefore close with the following open questions:

  1. Can we find a precise condition on the tiling that allows us to predict exactly when a free-form Sudoku graph is integral or not?

  2. If a given Sudoku is integral, is its blown up version always integral?

  3. Can a blown up free-form Sudoku be integral although its original Sudoku is not?

References

  • [1]

    Elsholtz C., Mütze A., Sudoku im Mathematikunterricht, Math. Semesterber 2007, 54, 69–93. CrossrefGoogle Scholar

  • [2]

    Evans R., Lindner B., Shi Y., Generating sudoku puzzles and its applications in teaching mathematics, Int. J. Math. Educ. Sci. Technol 2011, 42, 697–704. CrossrefGoogle Scholar

  • [3]

    Rosenhouse J., Taalman L., Taking Sudoku seriously. The math behind the world’s most popular pencil puzzle 2011, Oxford: Oxford University Press. Google Scholar

  • [4]

    Chen H., Cooper C., Solving sudoku: structures and strategies, Missouri J. Math. Sci 2017, 29, 12–18. Google Scholar

  • [5]

    Cameron P.J., Hilton A.J.W., Vaughan E.R., An analogue of Ryser’s theorem for partial Sudoku squares, J. Comb. Math. Comb. Comput 2012, 80, 47–69. Google Scholar

  • [6]

    Deng X.Q., Da Li Y., A novel hybrid genetic algorithm for solving Sudoku puzzles, Optim. Lett 2013, 7, 241–257. Web of ScienceCrossrefGoogle Scholar

  • [7]

    D’haeseleer J., Metsch K., Storme L., Van de Voorde G., On the maximality of a set of mutually orthogonal sudoku Latin squares, Des. Codes Cryptography 2017, 84, 143–152. CrossrefGoogle Scholar

  • [8]

    Lorch J., Constructing ordered orthogonal arrays via sudoku, J. Algebra Appl 2016, 15, 19 pages. Web of ScienceGoogle Scholar

  • [9]

    Fontana R., Random Latin squares and Sudoku designs generation, Electron. J. Stat 2014, 8, 883–893. CrossrefWeb of ScienceGoogle Scholar

  • [10]

    Keedwell A.D., Confirmation of a conjecture concerning orthogonal Sudoku and bimagic squares, Bull. Inst. Comb. Appl 2011, 63, 39–47. Google Scholar

  • [11]

    Arnold E., Field R., Lorch J., Lucas S., Taalman L., Nest graphs and minimal complete symmetry groups for magic Sudoku variants, Rocky Mt. J. Math 2015, 45, 887–901. CrossrefWeb of ScienceGoogle Scholar

  • [12]

    Smith J.D., Palindromic and sūdoku quasigroups, J. Comb. Math. Comb. Comput 2014, 88, 85–94. Google Scholar

  • [13]

    Jones S.K., Perkins S., Roach P.A., The structure of reduced Sudoku grids and the Sudoku symmetry group, Int. J. Comb 2012, 6 pages. Google Scholar

  • [14]

    Klotz W., Sander T., Wie kommt Sudoku zu ganzzahligen Eigenwerten?, Math. Semesterber 2010, 57, 169–183. CrossrefGoogle Scholar

  • [15]

    Sander T., Sudoku graphs are integral, Electron. J. Comb 2009, 16, research paper N25, 7 pages. Google Scholar

  • [16]

    Cvetkovic D.M., Petric M.V., Connectedness of the non-complete extended p-sum of graphs, Zb. Rad., Prir.-Mat. Fak., Univ. Novom Sadu, Ser. Mat 1983, 13, 345–352. Google Scholar

  • [17]

    Delorme C., Eigenvalues of complete multipartite graphs, Discrete Math 2012, 312, 2532–2535. CrossrefWeb of ScienceGoogle Scholar

  • [18]

    Horn R.A., Johnson C.R., Topics in Matrix Analysis 1991, Cambridge: Cambridge University Press. Google Scholar

  • [19]

    Cvetkovic D.M., Doob M., Sachs H., Spectra of graphs. Theory and applications. (3rd rev. a. enl. ed.) 1995, Leipzig: J. A. Barth Verlag. Google Scholar

About the article

Received: 2018-08-30

Accepted: 2018-11-26

Published Online: 2018-12-31


Citation Information: Open Mathematics, Volume 16, Issue 1, Pages 1445–1454, ISSN (Online) 2391-5455, DOI: https://doi.org/10.1515/math-2018-0125.

Export Citation

© 2018 Abudayah et al., published by De Gruyter. This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. BY-NC-ND 4.0

Comments (0)

Please log in or register to comment.
Log in