A supercomputer is a high-level performance computer in comparison to a general-purpose computer. Supercomputers are used for computationally intensive tasks in various fields: quantum mechanics, weather forecasting, climate research, molecular modeling, physical simulations, and much more.

Supercomputers were first developed in the 1960s, and Seymour Cray’s Control Data Corporation (CDC), Cray Research, and subsequent companies bearing his name or monogram were the fastest for decades. The earliest machines of this type were finely tuned conventional designs that outperformed their more general-purpose contemporaries. Increasing quantities of parallelism were introduced during the decade, with one to four processors being usual. Vector processors that operate on massive data arrays were popular in the 1970s. The Cray-1, which was released in 1976 and was a huge success, is a great example. Throughout the 1990s, vector computers were the most popular design.

Discovery is a supercomputer located at New Mexico State University, available to researchers around NMSU for free. Discovery has a high level of computing performance compared to a general-purpose computer.


A classic cluster is essentially computers grouped together in a manner that allows them to share infrastructure, such as disk space, and work together by sharing program data while those programs are running. However, this simple definition though accurate doesn’t capture the full capability of a modern cluster system, as it excludes an important idea. This idea, which has been developed to essentially become the core of clustering in general, is the scheduling system.

Scheduling System

The functional purpose of the scheduling system is to eliminate the need to know what individual computers are doing. A scheduling system aggregates data and monitors the system. A scheduler will keep an exact and up-to-date picture of what resources are available and where. Even beyond tracking resources, a scheduler will allow you to submit instructions for running your program. Then, it runs your program on your behalf once the necessary resources are available. The discovery Linux cluster consists of a login node, head node and compute nodes.