The general merge algorithm has a set of pointers p0..n that point to positions in a set of lists L0..n. Initially they point to the first item in each list. The algorithm is as follows:
While any of p0..n still point to data inside of L0..n instead of past the end:
- do something with the data items p0..n point to in their respective lists
- find out which of those pointers points to the item with the lowest key; advance one of those pointers to the next item in its list
Merge algorithms generally run in time proportional to the sum of the lengths of the lists; merge algorithms that operate on large numbers of lists at once will multiply the sum of the lengths of the lists by the time to figure out which of the pointers points to the lowest item, which can be accomplished with a heap-based priority queue in O(log n) time, for O(m log n) time, where n is the number of lists being merged and m is the sum of the lengths of the lists. When merging two lists of length m, there is a lower bound of 2m − 1 comparisons required in the worst case.
The classic merge (the one used in merge sort) outputs the data item with the lowest key at each step; given some sorted lists, it produces a sorted list containing all the elements in any of the input lists, and it does so in time proportional to the sum of the lengths of the input lists.
The C++'s Standard Template Library has the function
std::merge, which merges two sorted ranges of iterators, and
std::inplace_merge, which merges two consecutive sorted ranges in-place. In addition, the
std::list (linked list) class has its own
merge method which merges another list into itself. The type of the elements merged must support the less-than (<) operator, or it must be provided with a custom comparator.
Parallel merge can also be implemented using a divide-and-conquer algorithm, developed and shown in pseudo-code in. This algorithm performs well when combined with a fast sequential merge as a base case for merging of small arrays. Implementation using Intel's Threading Building Blocks (TBB) and Microsoft's Parallel Pattern Library (PPL) to run on multi-core processors is shown to perform well in practice.
- Berkman, Omer; Schieber, Baruch; Vishkin, Uzi (1993), "Optimal double logarithmic parallel algorithms based on finding all nearest smaller values", Journal of Algorithms 14 (3): 344–370, doi:10.1006/jagm.1993.1018
- Cormen et al. 2009, p. 800
- V. J. Duvanenko, "Parallel Merge", Dr. Dobb's Journal, February 2011
- Donald Knuth. The Art of Computer Programming, Volume 3: Sorting and Searching, Third Edition. Addison-Wesley, 1997. ISBN 0-201-89685-0. Pages 158–160 of section 5.2.4: Sorting by Merging. Section 5.3.2: Minimum-Comparison Merging, pp. 197–207.
- Cormen, Thomas; Leiserson, Charles; Rivest, Ronald; Stein, Clifford (2009), "Section 27.3: Multithreaded merge sort", Introduction to Algorithms (Third ed.), MIT Press and McGraw-Hill, pp. 797–804, ISBN 978-0-262-03384-8