“Compute refers to applications and workloads that require a great deal of computation, necessitating sufficient resources to handle these computation demands in an efficient manner.”
Hadoop is an open source, Java-based software programming framework that supports the processing and storage of extremely large data sets in a distributed computing environment. It is part of the Apache project sponsored by the Apache Software Foundation and delivers high-speed processing with the ability to run nearly limitless concurrent tasks delivering a highly desirable cost to performance ratio.
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures.
Companies seeking to create greater value from their HADOOP implementation should consult with a NordStar representative and explore the options available.