Jump to content

Sudoku code

From Wikipedia, the free encyclopedia

Sudoku codes are non-linear forward error correcting codes following rules of sudoku puzzles designed for an erasure channel. Based on this model, the transmitter sends a sequence of all symbols of a solved sudoku. The receiver either receives a symbol correctly or an erasure symbol to indicate that the symbol was not received. The decoder gets a matrix with missing entries and uses the constraints of sudoku puzzles to reconstruct a limited amount of erased symbols.

Sudoku codes are not suitable for practical usage but are subject of research. Questions like the rate and error performance are still unknown for general dimensions.[1]

In a sudoku one can find missing information by using different techniques to reproduce the full puzzle. This method can be seen as decoding a sudoku coded message that is sent over an erasure channel where some symbols got erased. By using the sudoku rules the decoder can recover the missing information. Sudokus can be modeled as a probabilistic graphical model and thus methods from decoding low-density parity-check codes like belief propagation can be used.

Erasure channel model

[edit]
The channel model for a standard sudoku erasure channel with a mapping from the input to the channel output with the erasure symbol and erasure probability .

In the erasure channel model a symbol gets either transmitted correctly with probability or is erased with probability (see Figure \ref{fig:Sudoku3x3channel}). The channel introduces no errors, i.e. no channel input is changed to another symbol. The example in Figure \ref{fig:Sudoku3x3BSC} shows the transmission of a Sudoku code. 5 of the 9 symbols got erased by the channel. The decoder is still able to reconstruct the message, i.e. the whole puzzle.

Schema of a sudoku transmission in the erasure channel model

Note that the symbols sent over the channel are not binary. For a binary channel the symbols (e.g. integers ) have to be mapped onto base 2. The binary erasure channel model however is not applicable because it erases only individual bits with some probability and not Sudoku symbols. If the symbols of the Sudoku are sent in packets the channel can be described as a packet erasure channel model.

Puzzle description

[edit]

A sudoku is a number-placement puzzle. It is filled in a way, that in each column, row and sub-grid N distinct symbols occur exactly once. Typical alphabet is the set of the integers . The size of the sub-grids limit the size of SUDOKUs to with . Every solved sudoku and every sub-grid of it is a Latin square, meaning every symbol occurs exactly once in each row and column. At the starting point (in this case after the erasure channel) the puzzle is only partially complete but has only one unique solution.

For channel codes also other varieties of sudokus are conceivable. Diagonal regions instead of square sub-grids can be used for performance investigations.[2] The diagonal sudoku has the advantage, that its size can be chosen more freely. Due to the sub-grid structure normal sudokus can only be of size n², diagonal sudokus have valid solutions for all odd .[2]

Sudoku codes are non-linear. In a linear code any linear combination of codewords give a new valid codeword, this does not hold for sudoku codes. The symbols of a sudoku are from a finite alphabet (e.g. integers ). The constraints of Sudoku codes are non-linear: all symbols within a constraint (row, line, sub-grid) must be different from any other symbol within this constraint. Hence there is no all-zero codeword in Sudoku codes.

Sudoku codes can be represented by probabilistic graphical model in which they take the form of a low-density parity-check code.[3]

Decoding with belief propagation

[edit]
Tanner graph of a Sudoku. denotes the entries of the Sudoku in row-scan order. denotes the constraint functions: associated with rows, associated with columns and associated with the sub-grids of the Sudoku.

There are several possible decoding methods for sudoku codes. Some algorithms are very specific developments for Sudoku codes. Several methods are described in sudoku solving algorithms. Another efficient method is with dancing links.

Decoding methods like belief propagation are also used for low-density parity-check codes are of special interest. Performance analysis of these methods on sudoku codes can help to understand decoding problems for low-density parity-check codes better.[3]

By modeling sudoku codes as a probabilistic graphical model belief propagation can be used for Sudoku codes. Belief propagation on the tanner graph or factor graph to decode Sudoku codes is discussed in by Sayir.[1] and Moon[4] This method is originally designed for low-density parity-check codes. Due to its generality belief propagation works not only with the classical Sudoku but with a variety of those. LDPC decoding is a common use-case for belief propagation, with slight modifications this approach can be used for solving Sudoku codes.[4]

The constraint satisfaction using a tanner graph is shown in the figure on the right. denotes the entries of the sudoku in row-scan order. denotes the constraint functions: associated with rows, associated with columns and associated with the sub-grids of the Sudoku. is defined as

[4]

Every cell is connected to 3 constraints: the row, column and sub-grid constraints. A specification of the general approach for belief propagation is suggested by Sayir:[1] The initial probability of a received symbol is either 1 to the observed symbol and 0 to all others or uniform distributed on the whole alphabet if the symbol is erased. For the belief propagation algorithm it is sufficient to transmit only a subset of possibilities instead of distributions, since the distribution is always uniform over the subset. The candidates for the erased symbols narrow down to a subset of the alphabet as symbols get excluded due to constraints. All values that are used by another cell in the constraint, and pairs that are shared among two other cells and so on are eliminated. Sudoku players use this method of logic excluding to solve most sudoku puzzles.

Encoding

[edit]

The aim of error-correcting codes is to encode data in a way such that it is more resilient to errors in the transmission process. The encoder has to map data to a valid sudoku grid from which the codeword can we taken e.g. in row-scan order.

shows the necessary steps.

A standard sudoku has about 72.5 bits of information as calculated in the next section. Information after Shannon is the degree of randomness in a set of data. An ideal coin toss for example has an information of bit. To represent the outcome of 72 coin tosses 72 bits are necessary. One Sudoku contains therewith about the same information as 72 coin tosses or a sequence of 72 bits. A sequence of 81 random symbols has bits of information. One Sudoku code can be seen as 72.5 bits of information and 184.3 bits redundancy. Theoretically a string of 72 bits can be mapped to one sudoku that is sent over the channel as a string of 81 symbols. However, there is no linear function that maps a string to a sudoku code.

A suggested encoding approach by Sayir[5] is as follows:

  • Start with an empty grid
  • Do the following for all entries sequentially
  • Use belief propagation to determine all valid symbols for the entry
  • If the cardinality of valid symbols is k>1 then convert the source randomness into a k-ary symbol and use it in the cell
Encoding example of a Sudoku with information and rate calculation.

For a sudoku the first entry can be filled with a source of cardinality 4. In this example this a 1. For the rest of this row, column and sub-grid this number is excluded from the possibilities in the belief propagation decoder. For the second cell only the numbers 2,3,4 are valid. The source has to be transformed into a uniform distribution between three possibilities and mapped to the valid numbers and so on, until the grid is filled.

Performance of sudoku codes

[edit]

The calculation of the rate of sudoku codes is not trivial. An example rate calculation of a sudoku is shown above. Filling line by line from the top left corner only the first entry has maximum information of bits. Every next entry cannot be any of the numbers used before, so the information reduces to , and for the following entries, as they have to be of the remaining left numbers. In the second lines the information is additionally reduced by the area rule: cell in row-scan order can only be a or as the numbers and are already used in the sub-grid. The last row contains no information at all. Adding all information up one gets bits. The rate in this example is

.

The exact number of possible Sudoku grids according to Mathematics of Sudoku is . With the total information of

the average rate of a standard Sudoku is

.

The average number of possible entries for a cell is or of information per Sudoku cell. Note that the rate may vary between codewords.[5]

The minimum number of given entries that render a unique solution was proven to be 17.[6] In the worst case only four missing entries can lead to ambiguous solutions. For an erasure channel it is very unlikely that 17 successful transmissions are enough to reproduce the puzzle. There are only about 50,000 known solutions with 17 given entries.[7]

Density evolution

[edit]

Density evolution is a capacity analysis algorithm originally developed for low-density parity-check codes on belief propagation decoding.[8] Density evolution can also be applied to Sudoku-type constraints.[1] One important simplification used in density evolution on LDPC codes is the sufficiency to analyze only the all-one codeword. With the Sudoku constraints however, this is not a valid codeword. Unlike for linear codes the weight-distance equivalence property does not hold for non-linear codes. Therewith it is necessary to compute density evolution recursions for every possible Sudoku puzzle to get precise performance analysis.

A proposed simplification is to analyze the probability distribution of the cardinalities of messages instead of the probability distribution of the message.[1] Density evolution is calculated on the entry nodes and the constraint nodes (compare Tanner graph above). On the entry nodes one analyzes the cardinalities of the constraints. If for example the constraints have the cardinalities then the entry can only be of one symbol. If the constraints have cardinalities then both constraints allow two different symbols. For both constraints the correct symbol is contained for sure, lets assume the correct symbol is . The other symbol can be equal or different for the constraints. If the symbols are different then the correct symbol is determined. If the second symbol is equal, lets assume the output symbols are of cardinality i.e. the symbols . Depending on the alphabet size () the probability for the unique output for the input cardinalities is

and for output of cardinality 2

For a standard Sudoku this results in a probability of for a unique solution. An analog calculation is done for all cardinality combinations. In the end the distribution of output cardinalities are summed up from the results. Note that the order of the input cardinality is interchangeable. The calculation of non-decreasing constraint combinations is therewith sufficient.

For constraint nodes the procedure is somewhat similar and described in the following example based on a standard Sudoku. Inputs to the constraint nodes are the possible symbols of the connected entry nodes. Cardinality 1 means that the symbol of the source node is already determined. Again a non-decreasing analysis is sufficient. Lets assume the true output value is 4 and the inputs have cardinalities with the true symbols 1-2-3. The messages with cardinality 1 are and . The message of cardinality 2 might be , or as the true symbol 3 must be contained. In two of three cases the output is the correct symbol 4 with cardinality 1: , , and , , . In one of three case the output cardinality is 2:, , . The output symbols in this case are . The final output cardinality distribution can be expressed by summing over all possible input combinations. For a standard Sudoku these are 64 combinations that can be grouped to 20 non-decreasing ones.[1]

If the cardinality converges to 1 the decoding is error-free. To find the threshold the erasure probability must be increased until the decoding error remains positive for any number of iterations. With the method of Sayir[1] density evolution recursions can be used to calculate thresholds also for Sudoku codes up to an alphabet size .

See also

[edit]

References

[edit]
  1. ^ a b c d e f g Sayir, Jossy; Atkins, Caroline (16 Jul 2014). "Density Evolution for SUDOKU codes on the Erasure Channel". Turbo Codes and Iterative Information Processing (ISTC), 2014 8th International Symposium on. arXiv:1407.4328. Bibcode:2014arXiv1407.4328A.
  2. ^ a b Sayir, Jossy (21 October 2014). "SUDOKU Codes, a class of non-linear iteratively decodable codes" (PDF). Retrieved 20 Dec 2015.
  3. ^ a b Khan, Sheehan; Jabbari, Shahab; Jabbari, Shahin; Ghanbarinejad, Majid. "Solving sudoku using probabilistic graphical models" (PDF). Retrieved 20 December 2015.
  4. ^ a b c Moon, T.K.; Gunther, J.H. (2006-07-01). "Multiple Constraint Satisfaction by Belief Propagation: An Example Using Sudoku". 2006 IEEE Mountain Workshop on Adaptive and Learning Systems. pp. 122–126. doi:10.1109/SMCALS.2006.250702. ISBN 978-1-4244-0166-6. S2CID 6131578.
  5. ^ a b Sayir, J.; Sarwar, J. (2015-06-01). "An investigation of SUDOKU-inspired non-linear codes with local constraints". 2015 IEEE International Symposium on Information Theory (ISIT). pp. 1921–1925. arXiv:1504.03946. doi:10.1109/ISIT.2015.7282790. ISBN 978-1-4673-7704-1. S2CID 5893535.
  6. ^ McGuire, Gary; Tugemann, Bastian; Civario, Gilles (2012-01-01). "There is no 16-Clue Sudoku: Solving the Sudoku Minimum Number of Clues Problem". arXiv:1201.0749 [cs.DS].
  7. ^ "Minimum Sudoku". staffhome.ecm.uwa.edu.au. Retrieved 2015-12-20.
  8. ^ Chung, Sae-Young; Richardson, T.J.; Urbanke, R.L. (2001-02-01). "Analysis of sum-product decoding of low-density parity-check codes using a Gaussian approximation". IEEE Transactions on Information Theory. 47 (2): 657–670. CiteSeerX 10.1.1.106.7729. doi:10.1109/18.910580. ISSN 0018-9448.