Skip to main content

IBM Advances its High Performance Data Analytics Arsenal with its Spectrum Computing Platform

As the need for gathering data continues, organizations keep dealing with increasing amounts of information that need to be stored, processed, and analyzed faster and better, stimulating the growth and evolution of the high performance computing (HPC) market.

One key segment in this market the continues to grow, especially in recent years, is the high performance data analytics (HPDA) as organizations continue to adopt and evolve their big data and data lake initiatives, to the point that IDC forecasts that in 2018, the HPDA server market will reach $2.6 billion (23.5% CAGR) and the HPDA external storage market will add $1.6 billion (26.5% CAGR).

Is not strange then, that software companies like IBM are keen to develop solutions capable to address the specific HPDA market segment, and IBM haws been working specifically on it with its Spectrum software line.

IBM Evolves Spectrum to Keep the Pace with the HPDA Market Segment

Late in November of last year, IBM announced a brand new software offering called IBM Spectrum Computing. The announcement signals how IBM is setting itself on route to address a not huge but important market segment of the IT industry: The provision of analytics solutions for high performance workloads or High Performance Data Analytics (HPDA).

With this, IBM aims to continue its strategy to serve and provide advanced data management and analytics solutions for an increasing set of computing ecosystems and environments. The new offering will enable organizations to work on high data volumes being analyze within advanced analytics software solutions including Spark, TensorFlow or Caffé to deploy applications that use machine learning, artificial intelligence and deep learning.

As IBM mentions:

IBM Spectrum Computing uses intelligent workload and policy-driven resource management to optimize resources across the data center, on premises and in the cloud.
Also, according to IBM, the new software platforms is designed to set up distributed, mission-critical, HPC infrastructures, scalable to over 160,000 cores and enable execution of big data analytics applications at speeds up to 150x faster.

The new Spectrum Computing offering from IBM is now set to deliver open offerings developed especially to increase speed of adoption, and production of parallel processing and clustered computing tasks to allow organizations to:
  • Simplify deployment via cluster virtualization and enabling disperse systems to work together as one by including shared computing, data services and cluster management specifically configured deep learning/machine learning tasks.
  • Deliver artificial intelligence (AI) for the enterprise, enable centralized management and reporting and allow multi-tenant access with end-to-end security.
  • Openness to the latest technology so it can provide support for the latest open source tools and protocols, including Spark and other deep learning software options.
  • Ease the deployment and adoption of cognitive workloads so end users will have access to the consumption of cluster resources across applications without having the need for specialized cluster knowledge.
  • Elasticity for hybrid clouds by simplifying cloud usage in distributed clustered environments, incorporating cloud automated workload-driven provisioning and de-provisioning routines to encourage intelligent workload and data transfer to and from the cloud

IBM Spectrum Cluster Foundation Screencap (Courtesy of IBM)

According to the "big blue", the Spectrum Computing family has been designed to provide support for hybrid architectures so it can accommodate to specific workloads as well as to specific platforms to enable optimal performance. Interestingly, IBM Spectrum Computing offers support for mixed x86-IBM POWER platforms environments.

Scalable to over 160,000 cores, IBM aims to enable users to fulfill their high performance analytics deployment needs with state-of-the-art software-defined computing technology for complex distributed, mission-critical, high-performance computing (HPC) infrastructures.

Core components of the Spectrum Computing Platform include:
  • IBM Spectrum LSF. Spectrum platforms’ workload management solution which provides user management and workload administration tasks.
  • IBM Spectrum MPI. Spectrum’s message passing interface (MPI) with high-performance capabilities and configured specifically to support distributed computing environments.
  • IBM Spectrum Cluster Foundation. IBM Spectrum’s infrastructure life cycle management solution for scale-out environments and coming at no charge.
  • IBM High Performance Services. Hosted on the Softlayer (IBM Cloud), this is IBM’s HPC and storage clusters as a service infrastructure.
Additional to the launch is the coming of a new set of tools developed to help users to design artificial intelligence (AI) models using much of the leading deep learning frameworks including Tensor Flow and Caffé.

One key aspect of this set of releases includes the latest version of IBM Spectrum Scale software ―Spectrum’s advanced storage management― which now provides full support for moving workloads including unified files, objects and HDFS right from where they´re storage to where they´re analyzed.

This latest version of IBM Spectrum LSF Suites is now generally available, along with the new versions of IBM Spectrum Conductor with Spark, Deep Learning Impact and IBM Spectrum Scale.

So What?

Well, as solutions for HPDA evolve, so they market reach, we are seeing continues adoption within many industries and business segments: finance, security, retail, and the adoption of new technologies that will require solutions capable of analyzing petabytes of data through parallel processing compute resources.

The latest version of IBM Spectrum Computing aims to put IBM in the forefront of the HPDA segment and better compete with other giants in this market like HPE and Hitachi. A competition field that will be interesting to keep an eye on in the coming years.

Finally, feel free to check this Spectrum intro video from IBM and of course, to let me know your thoughts right below.


  1. Prescient information investigation utilizes this data to channel current just as recorded information to distinguish examples and figure occasions and conditions that may happen later on at a specific time, given gave parameters. Data Analytics Course

  2. Colorado and Kentucky are on the forefront of this push with Best Full Spectrum Hemp Oil farms springing up throughout Colorado and test projects happening in Kentucky.

  3. The ideal mix of straightforward bookkeeping programming, master guidance, and incredible assistance from contracted guaranteed bookkeepers for consultants !
    online accounting services


  4. I finally found great post here.I will get back here. I just added your blog to my bookmark sites. thanks.Quality posts is the crucial to invite the visitors to visit the web page, that's what this web page is science course in rohtak

  5. Best coupon destinations for coupon codes and promotion codes that get a good deal on almost any site. Utilize one of these coupon locaters before each Best Discount code site A very much spent get-away can clean up youthful personalities. summer vacation can be used for chasing after leisure activities and drawing in kids in sports. Break offers the best london travel guide from extraordinary lodgings and top attractions to roadtrips, motivation and master counsel. Visit Emporio Armani official site and investigate the new men's assortments by Emporio Armani: appeal, contemporaneity, and style. Marvelous sea shores, satiny warm oceans, lavish landscape, and interminable daylight - these are a portion of the top elements of the ideal tropical Beach Vacations


Post a Comment

Popular posts from this blog

Machine Learning and Cognitive Systems, Part 2: Big Data Analytics

In the first part of this series, I described a bit of what machine learning is and its potential to become a mainstream technology in the industry of enterprise software, and serve as the basis for many other advances in the incorporation of other technologies related to artificial intelligence and cognitive computing. I also mentioned briefly how machine language is becoming increasingly important for many companies in the business intelligence and analytics industry. In this post I will discuss further the importance that machine learning already has and can have in the analytics ecosystem, especially from a Big Data perspective. Machine learning in the context of BI and Big Data analytics Just as in the lab, and other areas, one of the reasons why machine learning became extremely important and useful in enterprise software is its potential to deal not just with huge amounts of data and extract knowledge from it—which can somehow be addressed with disciplines such as data

Next-generation Business Process Management (BPM)—Achieving Process Effectiveness, Pervasiveness, and Control

The range of what we think and do is limited by what we fail to notice. And because we fail to notice that we fail to notice there is little we can do to change until we notice how failing to notice shapes our thoughts and deeds. —R.D. Laing Amid the hype surrounding technology trends such as big data, cloud computing, or the Internet of Things, for a vast number of organizations, a quiet, persistent question remains unanswered: how do we ensure efficiency and control of our business operations? Business process efficiency and proficiency are essential ingredients for ensuring business growth and competitive advantage. Every day, organizations are discovering that their business process management (BPM) applications and practices are insufficient to take them to higher levels of effectiveness and control. Consumers of BPM technology are now pushing the limits of BPM practices, and BPM software providers are urging the technology forward. So what can we expect from the next

Teradata Open its Data Lake Management Strategy with Kylo: Literally

Still distilling good results from the acquisition of former consultancy company Think Big Analytics , Teradata , a powerhouse in the data management market took one step further to expand its data management stack and to make an interesting contribution to the open source community. Fully developed by the team at Think Big Analytics, in March of 2017 the company launched Kylo –a full data lake management solution– but with an interesting twist: as a contribution to the open source community. Offered as an open source project under the Apache 2.0 license Kylo is, according to Teradata, a new enterprise-ready data lake management platform that enables self-service data ingestion and preparation, as well the necessary functionality for managing metadata, governance and security. One appealing aspect of Kylo is it was developed over an eight year period, as the result of number of internal projects with Fortune 1000 customers which has enabled Teradata to incorporate several be