Transaction processing system
Transaction processing is a style of computing that divides work into individual, indivisible operations, called transactions. A transaction processing system (TPS) or transaction server is a software system, or software/hardware combination, that supports transaction processing .
- 1 History
- 2 List of transaction processing systems
- 3 Processing types
- 4 Transaction processing system features
- 5 Types of transaction processing
- 6 Databases for transaction processing
- 7 Backup procedures
- 8 See also
- 9 References
- 10 Further reading
The first transaction processing system was SABRE, made by IBM for American Airlines, which became operational in 1970. Designed to process up to 83,000 transactions a day, the system ran on two IBM 7090 computers. SABRE was migrated to IBM System/360 computers in 1972, and became an IBM product first as Airline control Program (ACP) and later as Transaction Processing Facility (TPF). In addition to airlines TPF is used by large banks, credit card companies, and hotel chains.
The Hewlett-Packard NonStop system (formerly Tandem NonStop) was a hardware and software system designed for Online Transaction Processing (OLTP) introduced in 1976. The systems were designed for transaction processing and provided an extreme level of availability and data integrit
List of transaction processing systems
- IBM Transaction Processing Facility (TPF) – 1960. At Amity Unlike most other transaction processing systems TPF is a dedicated operating system for transaction processing on IBM System z mainframes. Originally Airline Control Program (ACP).
- IBM Information Management System (IMS) – 1966. A joint hierarchical database and information management system with extensive transaction processing capabilities. Runs on OS/360 and successors.
- IBM Customer Information Control System (CICS) – 1969. A transaction manager designed for rapid, high-volume online processing, CICS originally used standard system datasets, but now has a connection to IBM's DB/2 relational database system. Runs on OS/360 and successors and DOS/360 and successors, IBM AIX, VM, and OS/2. Non-mainframe versions are called TXSeries.
- Tuxedo – 1980s. Transactions for Unix, Extended for Distributed Operations developed by AT&T Corporation, now owned by Oracle Corporation. Tuxedo is a cross-platform TPS.
- UNIVAC Transaction Interface Package (TIP) – 1970s. A transaction processing monitor for UNIVAC 1100/2200 series computers.
- Burroughs Corporation supported transaction processing capabilities in its MCP operating systems using GEMCOS (Generalized Message Control System of 1980). As of 2012 UNISYS ClearPath Enterprise Servers include Transaction Server, "an extremely flexible, high-performance message and application control system."
- Digital Equipment Corporation (DEC) Application Control and Management System (ACMS) – 1985. "Provides an environment for creating and controlling online transaction processing (OLTP) applications on the VMS operating system." Runs on VAX/VMS systems.
- Digital Equipment Corporation (DEC) Message Control System (MCS-10) for PDP-10 TOPS-10 systems.
- Honeywell Multics Transaction Processing. Feature (TP) – 1979.
- Transaction Management eXecutive (TMX) was NCR Corporation's proprietary transaction processing system running on NCR Tower 5000-series systems. This system was used mainly by financial institutions in the 1980s and 1990s.
- Hewlett-Packard NonStop system – 1976. NonStop is an integrated hardware and software system specifically designed for transaction processing. Originally from Tandem Computers.
- Transarc Encina – 1991. Transarc was purchased by IBM in 1994. Encina was discontinued as a product and folded into IBM's TXSeries. Encina support was discontinued in 2006.
Batch processing is execution of a series of programs (jobs) on a computer without manual intervention. Several transactions, called a batch are collected and processed at the same time. The results of each transaction are not immediately available when the transaction is being entered; there is a time delay.
"Real time systems attempt to guarantee an appropriate response to a stimulus or request quickly enough to affect the conditions that caused the stimulus." Each transaction in real-time processing is unique; it is not part of a group of transactions.
Time sharing is the sharing of a computer system among multiple users, usually giving each user the illusion that they have exclusive control of the system. The users may be working on the same project or different projects, but there are usually few restrictions on the type of work each user is doing.
A Transaction Processing System (TPS) is a type of information system that collects, stores, modifies and retrieves the data transactions of an enterprise. Transaction processing systems also attempt to provide predictable response times to requests, although this is not as critical as for real-time systems. Rather than allowing the user to run arbitrary programs as time-sharing, transaction processing allows only predefined, structured transactions. Each transaction is usually short duration and the processing activity for each transaction is programmed in advance.
Transaction processing system features
The following features are considered important in evaluating transaction processing systems.
Fast performance with a rapid response time is critical. Transaction processing systems are usually measured by the number of transactions they can process in a given period of time.
The system must be available during the time period when the users are entering transactions. Many organizations rely heavily on their TPS; a breakdown will disrupt operations or even stop the business.
The system must be able to handle hardware or software problems without corrupting data. Multiple users must be protected from attempting to change the same piece of data at the same time, for example two operators cannot sell the same seat on an airplane.
Ease of use
Often users of transaction processing systems are casual users. The system should be simple for them to understand, protect them from data-entry errors as much as possible, and allow them to easily correct their errors.
The system should be capable of growth at incremental costs, rather than requiring a complete replacement. It should be possible to add, replace, or update hardware and software components without shutting down the system.
Types of transaction processing
Processing in a batch
Transactions may be collected and processed as in batch processing. Transactions will be collected and later updated as a batch when it's convenient or economical to process them. Historically, this was the most common method as the information technology did not exist to allow real-time processing.
Processing in real-time
This is the immediate processing of data. It provides instant confirmation of a transaction. It may involve a large number of users who are simultaneously performing transactions which change data. Because of advances in technology (such as the increase in the speed of data transmission and larger bandwidth), real-time updating is possible.
Databases for transaction processing
A database is an organized collection of data. Databases offer fast retrieval times for non-structured requests as in a typical transaction processing application.
Databases for transaction processing may be constructed using hierarchical, network, or relational structures.
- Hierarchical structure: organizes data in a series of levels. Its top to bottom like structure consists of nodes and branches; each child node has branches and is only linked to one higher level parent node.
- Network structure: network structures also organizes data using nodes and branches. But, unlike hierarchical, each child node can be linked to multiple, higher parent nodes.
- Relational structure: a relational database organizes its data in a series of related tables. This gives flexibility as relationships between the tables are built.
The following features are desirable in a database system used in transaction processing systems:
- Good data placement: The database should be designed to access patterns of data from many simultaneous users.
- Short transactions: Short transactions enables quick processing. This avoids concurrency and paces the systems.
- Real-time backup: Backup should be scheduled between low times of activity to prevent lag of the server.
- High normalization: This lowers redundant information to increase the speed and improve concurrency, this also improves backups.
- Archiving of historical data: Uncommonly used data are moved into other databases or backed up tables. This keeps tables small and also improves backup times.
- Good hardware configuration: Hardware must be able to handle many users and provide quick response times.
Since business organizations have become very dependent on transaction processing, a breakdown may disrupt the business' regular routine and stop its operation for a certain amount of time. In order to prevent data loss and minimize disruptions there have to be well-designed backup and recovery procedures. The recovery process can rebuild the system when it goes down.
A TPS may fail for many reasons such as system failure, human errors, hardware failure, incorrect or invalid data, computer viruses, software application errors or natural or man-made disasters. As it's not possible to prevent all failures, a TPS must be able to detect and correct errors when they occur and cope with failures. A TPS will go through a recovery of the database which may involve the backup, journal, checkpoint, and recovery manager:
- Journal: A journal maintains an audit trail of transactions and database changes. Transaction logs and Database change logs are used, a transaction log records all the essential data for each transactions, including data values, time of transaction and terminal number. A database change log contains before and after copies of records that have been modified by transactions.
- Checkpoint: The purpose of checkpointing is to provide a snapshot of the data within the database. A checkpoint, in general, is any identifier or other reference that identifies the state of the database at a point in time. Modifications to database pages are performed in memory and are not necessarily written to disk after every update. Therefore, periodically, the database system must perform a checkpoint to write these updates which are held in-memory to the storage disk. Writing these updates to storage disk creates a point in time in which the database system can apply changes contained in a transaction log during recovery after an unexpected shut down or crash of the database system. If a checkpoint is interrupted and a recovery is required, then the database system must start recovery from a previous successful checkpoint. Checkpointing can be either transaction-consistent or non-transaction-consistent (called also fuzzy checkpointing). Transaction-consistent checkpointing produces a persistent database image that is sufficient to recover the database to the state that was externally perceived at the moment of starting the checkpointing. A non-transaction-consistent checkpointing results in a persistent database image that is insufficient to perform a recovery of the database state. To perform the database recovery, additional information is needed, typically contained in transaction logs. Transaction consistent checkpointing refers to a consistent database, which doesn't necessarily include all the latest committed transactions, but all modifications made by transactions, that were committed at the time checkpoint creation was started, are fully present. A non-consistent transaction refers to a checkpoint which is not necessarily a consistent database, and can't be recovered to one without all log records generated for open transactions included in the checkpoint. Depending on the type of database management system implemented a checkpoint may incorporate indexes or storage pages (user data), indexes and storage pages. If no indexes are incorporated into the checkpoint, indexes must be created when the database is restored from the checkpoint image.
- Recovery Manager: A recovery manager is a program which restores the database to a correct condition which allows transaction processing to be restarted.
Depending on how the system failed, there can be two different recovery procedures used. Generally, the procedures involves restoring data that has been collected from a backup device and then running the transaction processing again. Two types of recovery are backward recovery and forward recovery:
- Backward recovery: used to undo unwanted changes to the database. It reverses the changes made by transactions which have been aborted.
- Forward recovery: it starts with a backup copy of the database. The transaction will then reprocess according to the transaction journal that occurred between the time the backup was made and the present time.
Types of back-up procedures
There are two main types of back-up procedures: grandfather-father-son and partial backups:
This procedure involves taking complete backups of all data at regular intervals – daily, weekly, monthly, or whatever is appropriate. Multiple generations of backup are retained, often three which gives rise to the name. The most recent backup is the son, the previous the father, and the oldest backup is the grandfather. This method is commonly used for a batch transaction processing system with a magnetic tape. If the system fails during a batch run, the master file is recreated by restoring the son backup and then restarting the batch. However, if the son backup fails, is corrupted or destroyed, then the previous generation of backup (the father) is used. Likewise, if that fails, then the generation of backup previous to the father (i.e. the grandfather) is required. Of course the older the generation, the more the data may be out of date. Organize only of records that have changed. For example, a full backup could be performed weekly, and partial backups taken nightly. Recovery using this scheme involves restoring the last full backup and then restoring all partial backups in order to produce an up-to-date database. This process is quicker than taking only complete backups, at the expense of longer recovery time.
Backup plus journal
This technique is also used in conjunction with regular complete backups. The master file is backed up at regular intervals. Completed transactions since the last backup are stored separately and are called journals, or journal files. The master file can be recreated by restoring the last complete backup and then reprocessing transactions from the journal files. This will produce the most up-to-date copy of the database, but recovery may take longer because of the time required to process a volume of journal records.
- Transaction processing
- Server (computing)
- Online transaction processing
- Customer Integrated System
- Data warehouse
- IBM Corporation. "CICS Transaction Server for z/OS, Version 3.2 Transaction processing". Retrieved Nov 12, 2012.
- "Terminals Help Manage Aluminum Firm's Production". Computerworld. July 26, 1976. Retrieved November 14, 2012.
- UNISYS Corporation (2012). Transaction Server for ClearPath MCP Configuration Guide (PDF).
- Digital Equipment Corporation (1989). VAX ACMS Guide to Creating Transaction Processing Applications.
- Bell, Gordon. "Digital Computing Timeline (1985)". Retrieved November 15, 2012.
- Van Vleck, Thomas. "Multics Glossary -T-". Retrieved November 15, 2012.
- Transarc. "Corporate Overview". Archived from the original on February 3, 1999. Retrieved November 16, 2012.
- IBM Corporation. "TXSeries for Multiplatforms". Retrieved November 16, 2012.
- Schuster, Stewart A. (June 15, 1981). "In Depth: Relational Data Base Management". Computerworld. Retrieved November 16, 2012.
- Gerhard Weikum, Gottfried Vossen, Transactional information systems: theory, algorithms, and the practice of concurrency control and recovery, Morgan Kaufmann, 2002, ISBN 1-55860-508-8