Data processing is, generally, "the collection and manipulation of items of data to produce meaningful information." In this sense it can be considered a subset of information processing, "the change (processing) of information in any manner detectable by an observer." [note 1]
The term Data Processing (DP) has also been used previously to refer to a department within an organization responsible for the operation of data processing applications.
Data processing functions
Data processing may involve various processes, including:
- Validation – Ensuring that supplied data is correct and relevant.
- Sorting – "arranging items in some sequence and/or in different sets."
- Summarization – reducing detail data to its main points.
- Aggregation – combining multiple pieces of data.
- Analysis – the "collection, organization, analysis, interpretation and presentation of data."
- Reporting – list detail or summary data or computed information.
- Classification – separates data into various categories.
The United States Census Bureau illustrates the evolution of data processing from manual through electronic procedures.
Manual data processing
Although widespread use of the term data processing dates only from the nineteen-fifties, data processing functions have been performed manually for millennia. For example, bookkeeping involves functions such as posting transactions and producing reports like the balance sheet and the cash flow statement. Completely manual methods were augmented by the application of mechanical or electronic calculators. A person whose job was to perform calculations manually or using a calculator was called a "computer."
The 1850 United States Census schedule was the first to gather data by individual rather than household. A number of questions could be answered by making a check in the appropriate box on the form. From 1850 through 1880 the Census Bureau employed "a system of tallying, which, by reason of the increasing number of combinations of classifications required, became increasingly complex. Only a limited number of combinations could be recorded in one tally, so it was necessary to handle the schedules 5 or 6 times, for as many independent tallies." "It took over 7 years to publish the results of the 1880 census" using manual processing methods.
Automatic data processing
The term automatic data processing was applied to operations performed by means of unit record equipment, such as Herman Hollerith's application of punched card equipment for the 1890 United States Census. "Using Hollerith's punchcard equipment, the Census Office was able to complete tabulating most of the 1890 census data in 2 to 3 years, compared with 7 to 8 years for the 1880 census.... It is also estimated that using Herman Hollerith's system saved some $5 million in processing costs" (in 1890 dollars) even with twice as many questions as during 1880.
Electronic data processing
Computerized data processing, or Electronic data processing represents a later development, with a computer used instead of several independent pieces of equipment. The Census Bureau first made limited use of electronic computers for the 1950 United States Census, using a UNIVAC I system, delivered in 1952.
The term data processing has mostly been subsumed by the newer and somewhat more general term information technology (IT). The term "data processing" is presently considered sometimes to have a negative connotation, suggesting use of older technologies. As an example, in 1996 the Data Processing Management Association (DPMA) changed its name to the Association of Information Technology Professionals. Nevertheless, the terms are approximately synonymous.
Commercial data processing
Commercial data processing involves a large volume of input data, relatively few computational operations, and a large volume of output. For example, an insurance company needs to keep records on tens or hundreds of thousands of policies, print and mail bills, and receive and post payments.
For science or engineering, the terms data processing and information systems are considered too broad, and the more specialized term data analysis is typically used. Data analysis uses specialized and precise algorithms and statistical calculations that are less often observed in a typical general business environment. For data analysis, software like SPSS or SAS, or their free counterparts such as DAP, gretl or PSPP are often used.
- French, Carl (1996). Data Processing and Information Technology (10th ed.). Thomson. p. 2. ISBN 1844801004.
- Illingworth, Valerie (11 December 1997). Dictionary of Computing. Oxford Paperback Reference (4th ed.). Oxford University Press. ISBN 9780192800466.
- Google N gram viewer. Retrieved June 26, 2013.
- Truesdell, Leon E. (1965). The development of punch card tabulation in the Bureau of the Census, 1890. United States Department of Commerce.
- Bohme, Frederick; Wyatt, J. Paul; Curry, James P. (1991). 100 Years of Data Processing: The Punchcard Century. United States Bureau of the Census.