Greatest common divisor

From Wikipedia, the free encyclopedia
Jump to: navigation, search
 \gcd(1,x) = y, or Thomae's function. Hatching at bottom indicates ellipses.

In mathematics, the greatest common divisor (gcd), also known as the greatest common factor (gcf), highest common factor (hcf), or greatest common measure (gcm), of two or more integers (at least one of which is not zero), is the largest positive integer that divides the numbers without a remainder. For example, the GCD of 8 and 12 is 4.[1][2]

This notion can be extended to polynomials, see Polynomial greatest common divisor, or to rational numbers (with integer quotients).

Overview[edit]

Notation[edit]

In this article we will denote the greatest common divisor of two integers a and b as gcd(a,b). Some older textbooks use (a,b).[1][2]

Example[edit]

The number 54 can be expressed as a product of two other integers in several different ways:

 54 \times 1 = 27 \times 2 = 18 \times 3 = 9 \times 6. \,

Thus the divisors of 54 are:

 1, 2, 3, 6, 9, 18, 27, 54. \,

Similarly the divisors of 24 are:

 1, 2, 3, 4, 6, 8, 12, 24. \,

The numbers that these two lists share in common are the common divisors of 54 and 24:

 1, 2, 3, 6. \,

The greatest of these is 6. That is the greatest common divisor of 54 and 24. One writes:

 \gcd(54,24) = 6. \,

Reducing fractions[edit]

The greatest common divisor is useful for reducing fractions to be in lowest terms. For example, gcd(42, 56) = 14, therefore,

\frac{42}{56}=\frac{3 \cdot 14 }{ 4 \cdot 14}=\frac{3 }{ 4}.

Coprime numbers[edit]

Two numbers are called relatively prime, or coprime, if their greatest common divisor equals 1. For example, 9 and 28 are relatively prime.

A geometric view[edit]

"Tall, slender rectangle divided into a grid of squares. The rectangle is two squares wide and five squares tall."
A 24-by-60 rectangle is covered with ten 12-by-12 square tiles, where 12 is the GCD of 24 and 60. More generally, an a-by-b rectangle can be covered with square tiles of side-length c only if c is a common divisor of a and b.

For example, a 24-by-60 rectangular area can be divided into a grid of: 1-by-1 squares, 2-by-2 squares, 3-by-3 squares, 4-by-4 squares, 6-by-6 squares or 12-by-12 squares. Therefore, 12 is the greatest common divisor of 24 and 60. A 24-by-60 rectangular area can be divided into a grid of 12-by-12 squares, with two squares along one edge (24/12 = 2) and five squares along the other (60/12 = 5).

Calculation[edit]

Using prime factorizations[edit]

Greatest common divisors can in principle be computed by determining the prime factorizations of the two numbers and comparing factors, as in the following example: to compute gcd(18, 84), we find the prime factorizations 18 = 2 · 32 and 84 = 22 · 3 · 7 and notice that the "overlap" of the two expressions is 2 · 3; so gcd(18, 84) = 6. In practice, this method is only feasible for small numbers; computing prime factorizations in general takes far too long.

Here is another concrete example, illustrated by a Venn diagram. Suppose it is desired to find the greatest common divisor of 48 and 180. First, find the prime factorizations of the two numbers:

48 = 2 × 2 × 2 × 2 × 3,
180 = 2 × 2 × 3 × 3 × 5.

What they share in common is two "2"s and a "3":

Least common multiple.svg
Least common multiple = 2 × 2 × 2 × 2 × 3 × 3 × 5 = 720
Greatest common divisor = 2 × 2 × 3 = 12.

Using Euclid's algorithm[edit]

A much more efficient method is the Euclidean algorithm, which uses a division algorithm such as long division in combination with the observation that the gcd of two numbers also divides their difference. To compute gcd(48,18), divide 48 by 18 to get a quotient of 2 and a remainder of 12. Then divide 18 by 12 to get a quotient of 1 and a remainder of 6. Then divide 12 by 6 to get a remainder of 0, which means that 6 is the gcd. Note that we ignored the quotient in each step except to notice when the remainder reached 0, signalling that we had arrived at the answer. Formally the algorithm can be described as:

\gcd(a,0) = a
\gcd(a,b) = \gcd(b, a \,\mathrm{mod}\, b),

where

 a \,\mathrm{mod}\, b = a - b \left\lfloor {a \over b} \right\rfloor .

If the arguments are both greater than zero then the algorithm can be written in more elementary terms as follows:

\gcd(a,a) = a
\gcd(a,b) = \gcd(a - b,b)\quad, if a > b
\gcd(a,b) = \gcd(a, b-a)\quad, if b > a

Complexity of Euclidean method[edit]

The existence of the Euclidean algorithm places (the decision problem version of) the greatest common divisor problem in P, the class of problems solvable in polynomial time. The GCD problem is not known to be in NC, and so there is no known way to parallelize its computation across many processors; nor is it known to be P-complete, which would imply that it is unlikely to be possible to parallelize GCD computation. In this sense the GCD problem is analogous to e.g. the integer factorization problem, which has no known polynomial-time algorithm, but is not known to be NP-complete. Shallcross et al. showed that a related problem (EUGCD, determining the remainder sequence arising during the Euclidean algorithm) is NC-equivalent to the problem of integer linear programming with two variables; if either problem is in NC or is P-complete, the other is as well.[3] Since NC contains NL, it is also unknown whether a space-efficient algorithm for computing the GCD exists, even for nondeterministic Turing machines.

Although the problem is not known to be in NC, parallel algorithms with time superior to the Euclidean algorithm exist; the best known deterministic algorithm is by Chor and Goldreich, which (in the CRCW-PRAM model) can solve the problem in O(n/log n) time with n1+ε processors.[4] Randomized algorithms can solve the problem in O((log n)2) time on \exp\left[O\left(\sqrt{n \log n}\right)\right] processors (note this is superpolynomial).[5]

Binary method[edit]

An alternative method of computing the gcd is the binary gcd method which uses only subtraction and division by 2. In outline the method is as follows: Let a and b be the two non negative integers. Also set the integer d to 1. There are now four possibilities:

  • Both a and b are even.

In this case 2 is a common factor. Divide both a and b by 2, double d, and continue.

  • a is even and b is odd.

In this case 2 is not a common factor. Divide a by 2 and continue.

  • a is odd and b is even.

Like the previous case 2 is not a common factor. Divide b by 2 and continue.

  • Both a and b are odd.

Without loss of generality, assume that for a and b as they are now, ab. In this case let c = (ab)/2. Then gcd(a,b) = gcd(a,c) = gcd(b,c). Because ba it is usually easier (and computationally faster) to determine gcd(b,c). If computing this algorithm by hand, gcd(b,c) may be apparent. Otherwise continue the algorithm until c = 0. Note that the gcd of the original a and b is still d times larger than the gcd of the odd a and odd b above. For further details see Binary GCD algorithm.

Example: a = 48, b = 18, d = 1 → 24, 9, 2 → 12, 9, 2 → 6, 9, 2 → 3, 9, 2 → c = 3; since gcd(9,3) = 3, the gcd originally sought is d times larger, namely 6.

Other methods[edit]

If a and b are both nonzero, the greatest common divisor of a and b can be computed by using least common multiple (lcm) of a and b:

\gcd(a,b)=\frac{a\cdot b}{\operatorname{lcm}(a,b)},

but more commonly the lcm is computed from the gcd.

Using Thomae's function f,

\gcd(a,b) = a f\left(\frac b a\right),

which generalizes to a and b rational numbers or commensurable real numbers.

Keith Slavin has shown that for odd a ≥ 1:

\gcd(a,b)=\log_2\prod_{k=0}^{a-1} (1+e^{-2i\pi k b/a})

which is a function that can be evaluated for complex b.[6] Wolfgang Schramm has shown that

\gcd(a,b)=\sum\limits_{k=1}^a \exp (2\pi ikb/a) \cdot \sum\limits_{d\left| a\right.} \frac{c_d (k)}{d}

is an entire function in the variable b for all positive integers a where cd(k) is Ramanujan's sum.[7] Donald Knuth proved the following reduction:

\gcd(2^a-1, 2^b-1)=2^{\gcd(a,b)}-1

for non-negative integers a and b, where a and b are not both zero.[8] More generally

\gcd(n^a-1,n^b-1)=n^{\gcd(a,b)}-1 \,

which can be proven by considering the Euclidean algorithm in base n. Another useful identity relates \gcd(a,b) to the Euler's totient function:

 \gcd(a,b) = \sum_{k|a \; \hbox{and} \; k|b} \varphi(k).

Properties[edit]

  • Every common divisor of a and b is a divisor of gcd(a, b).
  • gcd(a, b), where a and b are not both zero, may be defined alternatively and equivalently as the smallest positive integer d which can be written in the form d = a·p + b·q, where p and q are integers. This expression is called Bézout's identity. Numbers p and q like this can be computed with the extended Euclidean algorithm.
  • gcd(a, 0) = |a|, for a ≠ 0, since any number is a divisor of 0, and the greatest divisor of a is |a|.[2] This is usually used as the base case in the Euclidean algorithm.
  • If a divides the product b·c, and gcd(a, b) = d, then a/d divides c.
  • If m is a non-negative integer, then gcd(m·a, m·b) = m·gcd(a, b).
  • If m is any integer, then gcd(a + m·b, b) = gcd(a, b).
  • If m is a nonzero common divisor of a and b, then gcd(a/m, b/m) = gcd(a, b)/m.
  • The gcd is a multiplicative function in the following sense: if a1 and a2 are relatively prime, then gcd(a1·a2, b) = gcd(a1, b)·gcd(a2, b). In particular, recalling that gcd is a positive integer valued function (i.e, gets natural values only) we obtain that gcd(a, b·c) = 1 if and only if gcd(a, b) = 1 and gcd(a, c) = 1.
  • The gcd is a commutative function: gcd(a, b) = gcd(b, a).
  • The gcd is an associative function: gcd(a, gcd(b, c)) = gcd(gcd(a, b), c).
  • The gcd of three numbers can be computed as gcd(a, b, c) = gcd(gcd(a, b), c), or in some different way by applying commutativity and associativity. This can be extended to any number of numbers.
  • gcd(a, b) is closely related to the least common multiple lcm(a, b): we have
gcd(a, b)·lcm(a, b) = a·b.
This formula is often used to compute least common multiples: one first computes the gcd with Euclid's algorithm and then divides the product of the given numbers by their gcd.
gcd(a, lcm(b, c)) = lcm(gcd(a, b), gcd(a, c))
lcm(a, gcd(b, c)) = gcd(lcm(a, b), lcm(a, c)).
  • It is sometimes useful to define gcd(0, 0) = 0 and lcm(0, 0) = 0 because then the natural numbers become a complete distributive lattice with gcd as meet and lcm as join operation. This extension of the definition is also compatible with the generalization for commutative rings given below.
  • In a Cartesian coordinate system, gcd(a, b) can be interpreted as the number of points with integral coordinates on the straight line joining the points (0, 0) and (a, b), excluding (0, 0).

Probabilities and expected value[edit]

In 1972, James E. Nymann showed that k integers, chosen independently and uniformly from {1,...,n}, are coprime with probability 1/ζ(k) as n goes to infinity.[9] (See coprime for a derivation.) This result was extended in 1987 to show that the probability that k random integers has greatest common divisor d is d-k/ζ(k).[10]

Using this information, the expected value of the greatest common divisor function can be seen (informally) to not exist when k = 2. In this case the probability that the gcd equals d is d−2/ζ(2), and since ζ(2) = π2/6 we have

\mathrm{E}( \mathrm{2} ) = \sum_{d=1}^\infty d \frac{6}{\pi^2 d^2} = \frac{6}{\pi^2} \sum_{d=1}^\infty \frac{1}{d}.

This last summation is the harmonic series, which diverges. However, when k ≥ 3, the expected value is well-defined, and by the above argument, it is

 \mathrm{E}(k) = \sum_{d=1}^\infty d^{1-k} \zeta(k)^{-1} = \frac{\zeta(k-1)}{\zeta(k)}.

For k = 3, this is approximately equal to 1.3684. For k = 4, it is approximately 1.1106.

The gcd in commutative rings[edit]

The notion of greatest common divisor can more generally be defined for elements of an arbitrary commutative ring, although in general there need not exist one for every pair of elements.

If R is a commutative ring, and a and b are in R, then an element d of R is called a common divisor of a and b if it divides both a and b (that is, if there are elements x and y in R such that d·x = a and d·y = b). If d is a common divisor of a and b, and every common divisor of a and b divides d, then d is called a greatest common divisor of a and b.

Note that with this definition, two elements a and b may very well have several greatest common divisors, or none at all. If R is an integral domain then any two gcd's of a and b must be associate elements, since by definition either one must divide the other; indeed if a gcd exists, any one of its associates is a gcd as well. Existence of a gcd is not assured in arbitrary integral domains. However if R is a unique factorization domain, then any two elements have a gcd, and more generally this is true in gcd domains. If R is a Euclidean domain in which euclidean division is given algorithmically (as is the case for instance when R = F[X] where F is a field, or when R is the ring of Gaussian integers), then greatest common divisors can be computed using a form of the Euclidean algorithm based on the division procedure.

The following is an example of an integral domain with two elements that do not have a gcd:

R = \mathbb{Z}\left[\sqrt{-3}\,\,\right],\quad a = 4 = 2\cdot 2 = \left(1+\sqrt{-3}\,\,\right)\left(1-\sqrt{-3}\,\,\right),\quad b = \left(1+\sqrt{-3}\,\,\right)\cdot 2.

The elements 2 and 1 + √(−3) are two "maximal common divisors" (i.e. any common divisor which is a multiple of 2 is associated to 2, the same holds for 1 + √(−3)), but they are not associated, so there is no greatest common divisor of a and b.

Corresponding to the Bézout property we may, in any commutative ring, consider the collection of elements of the form pa + qb, where p and q range over the ring. This is the ideal generated by a and b, and is denoted simply (ab). In a ring all of whose ideals are principal (a principal ideal domain or PID), this ideal will be identical with the set of multiples of some ring element d; then this d is a greatest common divisor of a and b. But the ideal (ab) can be useful even when there is no greatest common divisor of a and b. (Indeed, Ernst Kummer used this ideal as a replacement for a gcd in his treatment of Fermat's Last Theorem, although he envisioned it as the set of multiples of some hypothetical, or ideal, ring element d, whence the ring-theoretic term.)

See also[edit]

Notes[edit]

  1. ^ a b Long (1972, p. 33)
  2. ^ a b c Pettofrezzo & Byrkit (1970, p. 34)
  3. ^ Shallcross, D.; Pan, V.; Lin-Kriz, Y. (1993). "The NC equivalence of planar integer linear programming and Euclidean GCD". 34th IEEE Symp. Foundations of Computer Science. pp. 557–564. 
  4. ^ Chor, B.; Goldreich, O. (1990). "An improved parallel algorithm for integer GCD". Algorithmica 5 (1–4): 1–10. doi:10.1007/BF01840374. 
  5. ^ Adleman, L. M.; Kompella, K. (1988). "Using smoothness to achieve parallelism". 20th Annual ACM Symposium on Theory of Computing. New York. pp. 528–538. doi:10.1145/62212.62264. ISBN 0-89791-264-0. 
  6. ^ Slavin, Keith R. (2008). "Q-Binomials and the Greatest Common Divisor". Integers Electronic Journal of Combinatorial Number Theory (University of West Georgia, Charles University in Prague) 8: A5. Retrieved 2008-05-26. 
  7. ^ Schramm, Wolfgang (2008). "The Fourier transform of functions of the greatest common divisor". Integers Electronic Journal of Combinatorial Number Theory (University of West Georgia, Charles University in Prague) 8: A50. Retrieved 2008-11-25. 
  8. ^ Knuth, Donald E.; Graham, R. L.; Patashnik, O. (March 1994). Concrete Mathematics: A Foundation for Computer Science. Addison-Wesley. ISBN 0-201-55802-5. 
  9. ^ Nymann, J. E. (1972). "On the probability that k positive integers are relatively prime". Journal of Number Theory 4 (5): 469–473. doi:10.1016/0022-314X(72)90038-8. 
  10. ^ Chidambaraswamy, J.; Sitarmachandrarao, R. (1987). "On the probability that the values of m polynomials have a given g.c.d.". Journal of Number Theory 26 (3): 237–245. doi:10.1016/0022-314X(87)90081-3. 

References[edit]

Further reading[edit]

External links[edit]