Digital Edition

SYS-CON.TV
Does Big Data Need an HPC Boost?
What are firms doing with HPC?

This post is sponsored by The Business Value Exchange and HP Enterprise Services

When should High-Performance Computing (HPC) be considered an integral and essential part of the so-called ‘business transformation' process? This is not a question often posed.

We normally center our fascination with the business transformation process around any given firm's path toward secure (but productive) enterprise mobile computing, Big Data analytics, cloud computing flexibility and new age workflow constructs that embrace social media and online intercommunication.

But how about plain old raw power and High-Performance Computing - or HPC as we normally call it, shouldn't this opportunity to turbo-charge also form part of our current transformation plans?

Advanced and Complex Applications
HPC is defined as a computing environment that employs the use of parallel processing for running what we will call "advanced and complex applications" in an effective and efficient manner. To be more precise and define the term exactly - it applies to any computing environment where the system has the capability to function above a teraflop (or 1012 floating-point operations per second) in its operation.

Although it is true to say that most HPC systems up until now have been tasked with performing compute jobs in fields including scientific research, molecular engineering and (for example) high-grade military uses... thing are changing.

HPC is used to perform tasks including data storage (and, of course, analysis) and what we used to call (and sometime still do) data mining. It will also be used for running complex simulation scenarios and also in deep mathematical calculations and for the visualization of complex data.

... and so today, with so many firms becoming increasingly heavily digitised, the argument to take HPC forward into a wider range of business applications now arises.

In terms of wider usage, HPC can be used to develop, test and redesign products at the same time as optimizing production and delivery processes. Ultimately, Big Data will need HPC in order to be able to store, analyze and produce insight. HPC can also be used (alongside Big Data intelligence) to execute customer trend monitoring, searching and/or profiling.

What Are Firms Doing with HPC?
HP has announced that Airbus has boosted its HPC capacity for aircraft development to 1,200 Teraflops by deploying a new generation of HP Performance Optimized Datacenters (PODs) in Toulouse and Hamburg.

Each of Airbus' 12-meter-long containerized HP PODs delivers the equivalent of nearly 500 square meters of data center space and contains all the elements of an HP Converged Infrastructure: blade servers, storage, networking, software and management (as well as integrated power and cooling).

"Organizations like Airbus need creative scenarios to cater for future business needs," said Peter Ryan, senior vice president & general manager, HP Enterprise Group EMEA. "HP will continue to provide the newest, most powerful technology and operations to support Airbus' HPC for the next five years."

Big Data needs HPC and HPC needs to work on Big Data - this is a marriage made in heaven.

About Adrian Bridgwater
Adrian Bridgwater is a freelance journalist and corporate content creation specialist focusing on cross platform software application development as well as all related aspects software engineering, project management and technology as a whole.

In order to post a comment you need to be registered and logged in.

Register | Sign-in

Reader Feedback: Page 1 of 1



ADS BY GOOGLE
Subscribe to the World's Most Powerful Newsletters

ADS BY GOOGLE

Cloud-Native thinking and Serverless Computing are now the norm in financial services, manufacturing...
Most modern computer languages embed a lot of metadata in their application. We show how this goldmi...
Moving to Azure is the path to digital transformation, but not every journey is effective. Organizat...
Public clouds dominate IT conversations but the next phase of cloud evolutions are "multi" hybrid cl...
On-premise or off, you have powerful tools available to maximize the value of your infrastructure an...
At CloudEXPO Silicon Valley, June 24-26, 2019, Digital Transformation (DX) is a major focus with exp...
Every organization is facing their own Digital Transformation as they attempt to stay ahead of the c...
Data center, on-premise, public-cloud, private-cloud, multi-cloud, hybrid-cloud, IoT, AI, edge, SaaS...
DevOps has long focused on reinventing the SDLC (e.g. with CI/CD, ARA, pipeline automation etc.), wh...
Now is the time for a truly global DX event, to bring together the leading minds from the technology...
In a recent survey, Sumo Logic surveyed 1,500 customers who employ cloud services such as Amazon Web...
Atmosera delivers modern cloud services that maximize the advantages of cloud-based infrastructures....
In today's always-on world, customer expectations have changed. Competitive differentiation is deliv...
Artifex Software began 25-years ago with Ghostscript, a page description language (PDL) interpreter ...
In an age of borderless networks, security for the cloud and security for the corporate network can ...
Isomorphic Software is the global leader in high-end, web-based business applications. We develop, m...
Cloud Storage 2.0 has brought many innovations, including the availability of cloud storage services...
In very short order, the term "Blockchain" has lost an incredible amount of meaning. With too many j...
For enterprises to maintain business competitiveness in the digital economy, IT modernization is req...
At CloudEXPO Silicon Valley, June 24-26, 2019, Digital Transformation (DX) is a major focus with exp...