Monday, March 3, 2008

A Brief Introducation to Mainframes


INTRODUCTION:-

Mainframes (often colloquially referred to as Big Iron) are computers used mainly by large organizations for critical applications, typically bulk data processing such as census, industry and consumer statistics, ERP, and financial transaction processing.
The term probably originated from the early mainframes, as they were housed in enormous, room-sized metal boxes or frames.
[1] Later the term was used to distinguish high-end commercial machines from less powerful units which were often contained in smaller packages.
Today in practice, the term usually refers to computers compatible with the
IBM System/360 line, first introduced in 1965. (IBM System z10 is IBM's latest incarnation.) Otherwise, systems with similar functionality but not based on the IBM System/360 are referred to as "servers." However, "server" and "mainframe" are not synonymous (see client-server).
Some non-System/360-compatible systems derived from or compatible with older (pre-Web) server technology may also be considered mainframes. These include the
Burroughs large systems, the UNIVAC 1100/2200 series systems, and the pre-System/360 IBM 700/7000 series. Most large-scale computer system architectures were firmly established in the 1960s and most large computers were based on architecture established during that era up until the advent of Web servers in the 1990s. (Interestingly, the first Web server running anywhere outside Switzerland ran on an IBM mainframe at Stanford University as early as 1990. See History of the World Wide Web for details.)
There were several
minicomputer operating systems and architectures that arose in the 1970s and 1980s, but minicomputers are generally not considered mainframes. (UNIX arose as a minicomputer operating system; Unix has scaled up over the years to acquire some mainframe characteristics.)
Many defining characteristics of "mainframe" were established in the 1960s, but those characteristics continue to expand and evolve to the present day.

DESCRIPTION:-
Modern mainframe computers have abilities not so much defined by their single task computational speed (flops or clock rate) as by their redundant internal engineering and resulting high reliability and security, extensive input-output facilities, strict backward compatibility for older software, and high utilization rates to support massive throughput. These machines often run for years without interruption, with repairs and even software and hardware upgrades taking place during normal operation. More recently, there are several IBM mainframe installations that have delivered over a decade of continuous business service as of 2007, with upgrades not interrupting service. Mainframes are defined by high availability, one of the main reasons for their longevity, as they are used in applications where downtime would be costly or catastrophic. The term Reliability, Availability and Serviceability (RAS) is a defining characteristic of mainframe computers.
In the 1960s, most mainframes had no interactive interface. They accepted sets of
punch cards, paper tape, and/or magnetic tape and operated solely in batch mode to support back office functions, such as customer billing. Teletype devices were also common, at least for system operators. By the early 1970s, many mainframes acquired interactive user interfaces and operated as timesharing computers, supporting hundreds or thousands of users simultaneously along with batch processing. Users gained access through specialized terminals or, later, from personal computers equipped with terminal emulation software. Many mainframes supported graphical terminals (and terminal emulation) by the 1980s (if not earlier). Nowadays most mainframes have partially or entirely phased out classic user terminal access in favor of Web user interfaces.
Historically mainframes acquired their name in part because of their substantial size and requirements for specialized
HVAC and electrical power. Those requirements ended by the mid-1990s, with CMOS mainframe designs replacing the older bipolar technology. In fact, in a major reversal, IBM touts the mainframe's ability to reduce data center energy costs for power and cooling and reduced physical space requirements compared to server farms.

CHARACTERISTICS OF MAINFRAMES
Mainframes have the ability to run (or host) multiple operating systems and thereby operate not as a single computer but as a number of virtual machines. In this role, a single mainframe can replace dozens or even hundreds of smaller servers, reducing management and administrative costs while providing greatly improved scalability and reliability.
Mainframes can add or
hot swap system capacity non disruptively and granularly. Modern mainframes, notably the IBM zSeries and System z9 servers, offer three levels of virtualization: logical partitions (LPARs, via the PR/SM facility), virtual machines (via the z/VM operating system), and through its operating systems (notably z/OS with its key-protected address spaces and sophisticated goal-oriented workload scheduling,[clarify] but also Linux and Java). This virtualization is so thorough, so well established, and so reliable that most IBM mainframe customers run no more than two machines: one in their primary data center, and one in their backup data center—fully active, partially active, or on standby—in case there is a catastrophe affecting the first building. All test, development, training, and production workload for all applications and all databases can run on a single machine, except for extremely large demands where the capacity of one machine might be limiting. Such a two mainframe installation can support continuous business service, avoiding both planned and unplanned outages.
Mainframes are designed to handle very high volume input and output (I/O) and emphasize throughput computing. Since the mid-1960's, mainframe designs have included several subsidiary computers (called
channels or peripheral processors) which manage the I/O devices, leaving the CPU free to deal only with high-speed memory. It is common in mainframe shops to deal with massive databases and files. Giga-record or tera-record files are not unusual.[citation needed] Compared to a typical PC, mainframes commonly have hundreds to thousands of times as much data storage online, and can access it much faster.[citation needed]
Mainframe
return on investment (ROI), like any other computing platform, is dependent on its ability to scale, support mixed workloads, reduce labor costs, deliver uninterrupted service for critical business applications, and several other risk-adjusted cost factors. Some argue that the modern mainframe is not cost-effective. Hewlett-Packard and Dell unsurprisingly take that view at least at times, and so do a few independent analysts. Sun Microsystems used to take that view but, beginning in mid-2007, started promoting its new partnership with IBM, including probable support for the company's OpenSolaris operating system running on IBM mainframes. The general consensus (held by Gartner[citation needed] and other independent analysts) is that the modern mainframe often has unique value and superior cost-effectiveness, especially for large scale enterprise computing. In fact, Hewlett-Packard also continues to manufacture its own mainframe (arguably), the NonStop system originally created by Tandem. Logical partitioning is now found in many UNIX-based servers, and many vendors are promoting virtualization technologies, in many ways validating the mainframe's design accomplishments.
Mainframes also have unique execution integrity characteristics for
fault tolerant computing. System z9 servers execute each instruction twice,[citation needed] compare results, and shift workloads "in flight" to functioning processors, including spares, without any impact to applications or users. This feature, also found in HP's NonStop systems, is known as lock-stepping, because both processors take their "steps" (i.e. instructions) together. Not all applications absolutely need the assured integrity that these systems provide, but many do, such as financial transaction processing.
Despite these differences, the IBM mainframe, in particular, is still a general purpose business computer in terms of its support for a wide variety of popular operating systems,
middleware, and applications.


MAINFRAMES VS SUPERCOMPUTERS
The distinction between supercomputers and mainframes is not a hard and fast one, but supercomputers generally focus on problems which are limited by calculation speed while mainframes focus on problems which are limited by input/output and reliability ("throughput computing") and on solving multiple business problems concurrently (mixed workload). The differences and similarities include:

  1. Both types of systems offer parallel processing. Supercomputers typically expose it to the programmer in complex manners, while mainframes typically use it to run multiple tasks. One result of this difference is that adding processors to a mainframe often speeds up the entire workload transparently.
    Supercomputers are optimized for complicated computations that take place largely in memory, while mainframes are optimized for comparatively simple computations involving huge amounts of external data. For example, weather forecasting is suited to supercomputers, and insurance business or payroll processing applications are more suited to mainframes.
  2. Supercomputers are often purpose-built for one or a very few specific institutional tasks (e.g. simulation and modeling). Mainframes typically handle a wider variety of tasks (e.g. data processing, warehousing). Consequently, most supercomputers can be one-off designs, whereas mainframes typically form part of a manufacturer's standard model lineup.
  3. Mainframes tend to have numerous ancillary service processors assisting their main central processors (for cryptographic support, I/O handling, monitoring, memory handling, etc.) so that the actual "processor count" is much higher than would otherwise be obvious. Supercomputer design tends not to include as many service processors since they don't appreciably add to raw number-crunching power.
    There has been some blurring of the term "mainframe," with some PC and server vendors referring to their systems as "mainframes" or "mainframe-like." This is not widely accepted and the market generally recognizes that mainframes are genuinely and demonstrably different.


SPEED AND PERFORMANCE

The CPU speed of mainframes has historically been measured in millions of instructions per second (MIPS). MIPS have been used as an easy comparative rating of the speed and capacity of mainframes. The smallest System z9 IBM mainframes today run at about 26 MIPS and the largest about 17,801 MIPS. IBM's Parallel Sysplex technology can join up to 32 of these systems, making them behave like a single, logical computing facility of as much as about 569,632 MIPS.


The MIPS measurement has long been known to be misleading and has often been parodied as "Meaningless Indicator of Processor Speed." The complex CPU architectures of modern mainframes have reduced the relevance of MIPS ratings to the actual number of instructions executed. Likewise, the modern "balanced performance" system designs focus both on CPU power and on I/O capacity, and virtualization capabilities make comparative measurements even more difficult. See benchmark (computing) for a brief discussion of the difficulties in benchmarking such systems. IBM has long published a set of LSPR (Large System Performance Reference) ratio tables for mainframes that take into account different types of workloads and are a more representative measurement. However, these comparisons are not available for non-IBM systems. It takes a fair amount of work (and maybe guesswork) for users to determine what type of workload they have and then apply only the LSPR values most relevant to them.
To give some idea of real world experience, it is typical for a single mainframe CPU to execute the equivalent of 50, 100, or even more distributed processors' worth of activity, depending on the workloads. Merely counting processors to compare server platforms is extremely perilous.

HISTORY OF MAINFRAME COMPUTERS





1960's
Burroughs B5000 mainframe introduced. The system can be considered the first of the "third generation" of computer systems. The most remarked-upon aspects are its use of a hardware-managed stack for calculation, and the extensive use of descriptors for data access. It included virtual memory -- perhaps the first commercial computer to do so -- as well as support for multiprogramming and multiprocessing.

1964
CDC (Computer Data Corp.) 6600 shipped; 100 nsec cycle time.
First GE Time-sharing operation at Dartmouth College of the DTSS Dartmouth time-sharing system on a GE-265 (GE-225 + Datanet-30)

IBM announces the 360 family of computer systems.

The Burroughs B5500, appeared. It also had multiprogramming and virtual memory capabilities, but was three times faster than the B5000


1965
IBM ships the midrange 360 model 40 computer which had COBOL and FORTRAN programming languages available as well as the stock Basic Assembly Language (BAL) assembler.

Introduction of GECOS-II, a multi-programming operating system for the GE-600

1966
The Burroughs B6500, which was actually an improved version of the B5500.

1967

First IBM 360/Model 91 shipped to NASA GSFC.

1969
First shipment of the CDC 7600 computer system.
First shipment of IBM 360 Model 85. The 360 family was intended to have 3 operating systems:
DOS/360 operating system for the small machines. It could run two "real-time" sessions and one batch session.
OS/360 operating system for the midrange and high end.
TSS/360 operating system for Time-sharing Multi-user system

Introduction of Honeywell model 115 in the H-200 product line. The line was renamed H-2000 after models 115/2, 1015 and 2015 introduced in January 1971, and model 2020 and 2030 in December 1972 after the GE merger. The line was eventually merged into Series6 0 NPL through a H-200 mode (emulator) on level 64.

Introduction of GE-655 that was better known as H-6000 after 1970.
1970
Burroughs announces the 700 series. The first B6700 computer systems were installed during 1971. It was the first Burroughs machine with dynamic linking of programs at runtime. The B6700 line started out with one CPU and one i/o processor and could be expanded up to a maximum of three CPUs and two i/o processors.

Formal acquisition of Bull-General Electric by Honeywell. BGE takes the name of Honeywell-Bull.
IBM announces a family of machines with an enhanced instruction set, called System/370. The 370s proved so popular that there was a two-year waiting list of customers who had ordered a systems.
A giant dies: Announcement of the cession of the world-wide GE computer business, except time-sharing to Honeywell.

1971
US Air Force orders several Honeywell H-6000 WWMCCS (World Wide Military Command and Control System), a $3.5M contract.

First shipments of IBM S/370 Models 155 and 165 as well as the S/360 Model 195.
1973
Introduction of virtual memory on IBM S/370 Models 158 and 168.
1975
Amdahl 470 V/6 computer system delivered to NASA.
1977
The Burroughs Scientific Processor was developed, and announced.

IBM 3033 computer system announced
1979
The Burroughs 900-level systems were introduced.
1985
The most powerful IBM computer system of its time, the 3090 high-end processor of the IBM 308X computer series incorporated one-million-bit memory chips, Thermal Conduction Modules to provide the shortest average chip-to-chip communication time of any large general purpose computer. The Model 200 (entry-level with two central processors) and Model 400 (with four central processors) IBM 3090 had 64 and 128 megabytes of central storage, respectively. At the time of announcement, the purchase price of a Model 200 was $5 million. A later six-processor IBM 3090 Model 600E, using vector processors, could perform computations up to 14 times faster than the earlier four-processor IBM 3084.

1990
Th
e ES/9000 models came out with fiber-optical I/O channels (ESCON), and IBM began using the name System/390. The ES/9000s exploited new technologies, such as high-speed fiber optic channels with IBM's new ESCON architecture, ultra-dense circuits and circuit packaging that provided higher performance, extended supercomputing capabilities and twice the processor memory previously available. The line spanned a 100-fold performance range increase from the smallest (model 120) to the most powerful (model 900 six-way multiprocessor). Basic purchase prices for the air-cooled processors of ES/9000 ranged from approximately $70,500 to $3.12 million. Basic purchase prices for the water-cooled models ranged from $2.45 million to $22.8 million.
1999
IBM releases a new generation of S/390.
2002
The S/390 G5/G6 enterprise server family has up to 256 channels, from 2 to 8 Cryptographic Coprocessors, from 8 to 32 Gigabytes of memory, and can run under OS/390, MVS, VM, VSE, or TPF operating systems. It can also host an unbelievable amount of hard drive storage.
Let's see a PC match that!
2004
The 3/4 ton IBM eServer zSeries 890, dubbed the "Baby Shark" can host up to 32 GBytes of memory. The four PCIX Crypto Coprocessor (and optional PCI Crypto Accelerators) on the z890 have seven engine levels, giving a total of 28 capacity settings overall.
With it's advanced virtualization technology the 64-bit z890 can run several operating systems at the same time including z/OS, OS/390®, z/VM®, VM/ESA®, VSE/ESA, TPF and Linux for zSeries and Linux for S/390®.
The z890 is upgradeable within z890 family and can also upgrade to z990 from select z890 configurations.
Configured with the new Enterprise
Storage Server Model 750 which handles from 1.1TB up to 4.6TB of data, the x890 makes an awesome server.

2007
IBM produces the Blue-Gene/P, a system capable of a petaflop (1,000,000 gigaflops or 1,000 teraflops). It sports 73,728 processors comprised of four cores each of IBM’s 850MHz PowerPC 450, resulting in 294,912 cores. The system can be scaled to nearly three times that size, resulting in a 3 petaflop capability and is all hooked up via a high-end optical network.