VMblog 2017 Virtualization and Cloud Prediction
Contributed by George Teixeira, President and CEO, DataCore Software
Contributed by George Teixeira, President and CEO, DataCore Software
With so much computing power still sitting idle - despite all of the incredible technology advancements that have occurred - in 2017, the time is right for parallel processing software to go mainstream and unleash the immense processing power of today's multicore systems to positively disrupt the economic and productivity impact of what computing can do and where it can be applied.
New software innovations will make 2017 a breakout year for parallel processing. The key is that the software has to become simple to use and non-disruptive to applications to allow it to move from specialized use cases to general application usage. By doing so, the impact of this will be massive because application performance, enterprise workloads and greater consolidation densities on virtual platforms and in cloud computing that have been stifled by the growing gap between compute and I/O will no longer be held back. This will be realized with new parallel I/O software technologies now available that are easy to use, require no changes to the applications and are capable of fully leveraging the power of multicores to dramatically increase productivity and overcome the I/O bottleneck that has been holding back our industry; this is the catalyst of change.
Parallel processing software can now go beyond the realm of specialized uses such as HPC and areas like genomics that have focused primarily on computation, and impact the broader world of applications that require real-time responses and interactions. This includes mainstream applications and storage that drive business transactions, cloud computing, databases, data analytics, as well as the interactive worlds of machine learning and the Internet of Things (IoT).
The real driver of change is the economic and productivity disruption. Today, many new applications such as analytics are not practical because they require hundreds if not thousands of servers to get the job done; yet each server is becoming capable of supporting hundreds of multi-threading computing cores, all available to drive workloads that until now have sat there idle, waiting for work to do. We are ushering in an era where one server will do the work of 10 -- or 100 servers -- of the past. This will be the result of parallel processing software that unlocks the full utilization of multicores, leading to a revolution in productivity and making a new world of applications affordable to mainstream IT in 2017.
The Impact on Real-time Analytics and Big Data Performance will be Profound
The combination of faster response times and the multiplying impact on productivity through parallelization will fuel the next step forward in ‘real-time' analytics, big data and database performance. DataCore sees this as the next step forward in 2017. Our background in parallel processing, real-time I/O and software-defined storage has made our company uniquely well positioned to take advantage of the next big challenge in a world that requires the rate and amount of interactions and transactions to happen at a far faster pace with much faster response times.
The ability to do more work by doing it in parallel -- and to react quickly -- is the key. DataCore sees itself as helping to drive the step function change needed to make real-time analytics and big data performance practical and affordable. The implications on productivity and business decision making based on insights from data in areas such as financial, banking, retail, fraud detection, healthcare, and genomics, as well as machine learning and Internet of Things type applications, will be profound.
The Microsoft Impact Arrives: Azure Stack, Hybrid Cloud, Windows and SQL Server 2016
The success and growth of Microsoft's Azure Cloud has already become evident, however the real impact is the larger strategy of how Microsoft has worked to reconcile the world of on-premise and cloud computing. Microsoft was one of the first cloud vendors to recognize that the world is not just public clouds but that it will continue to be a mix of on-premise and cloud. Microsoft's Azure Stack continues to advance in making it seamless to get the benefits of cloud-like computing whether in the cloud or within a private cloud. It has become the model for hybrid cloud computing. Likewise, Microsoft continues to further integrate its Windows and server solutions to work more seamlessly with cloud capabilities.
While Windows and Azure get most of the attention, one of the most dramatic changes at Microsoft has been how it has reinvented and transformed its database offerings into a true big data and analytics platform for the future. It is time to take another look at SQL Server 2016; it is far more powerful and capable, and now deals with all types of data. As a platform, it is primed to work with Microsoft's large eco-system of marketplace partners, including DataCore with its parallel processing innovations, to redefine what is possible in the enterprise, the cloud, and with big data performance and real-time analytic use cases for traditional business applications, as well as new developing use cases in machine learning, cognitive computing and the Internet of Things.
Storage has Transformed; It's Servers + Software-Defined Infrastructure!
We are the midst of an inevitable and increasing trend in which servers are defining what storage is. Escalating this trend DataCore used parallel I/O software technologies to power off-the-shelf multicore servers to drive the world's fastest storage systems in terms of performance, lowest latencies and best price-performance. Traditional storage systems can no longer keep up and are on the decline, and as a result, are increasingly being replaced by commodity servers and software-defined infrastructure solutions that can leverage their power to solve the growing data storage problem. The storage function and associated data services are now being driven by software and becoming another "application workload" running on these cost-efficient server platforms, and this wave of flexible server-based storage systems are already having a disruptive industry impact.
Marketed as server-SANs, virtual SANs, web-scale, scale-out and hyper-converged systems, they are a collection of standard off-the-shelf servers, flash cards and disk drives - but it is the software that truly defines their value differentiation. Storage has become a server game. Parallel processing software and the ability to leverage multicore server technology is the major game-changer. In combination with software-defined infrastructure, it will lead to a productivity revolution and further solidify "servers as the new storage." For additional information, see the following report:http://wikibon.com/server-san-readies-for-enterprise-and-cloud-domination/
What's Beyond Flash?
Remember when flash was the next big thing? Now it's here. What is the next step -- how do we go faster and do more with less? The answer is obvious; if flash is now here and yet performance and productivity are still an issue for many enterprise applications especially database use cases, then we need to parallelize the I/O processing. Why? It multiplies what can be done as a result of many compute engines working in parallel to process and remove bottlenecks and queuing delays higher up in the stack, near the application, so we avoid as much device level I/O as possible and drive performance and response times far beyond any single device level optimization that flash/SSD alone can deliver. The power of the ‘many' far exceed what only ‘one' can do - combining flash and parallel I/O enables users to drive more applications faster, do more work and open up applications and use cases that have been previously impossible to do.
Going Beyond Hyper-Convergence: Hyper-Productivity is the Real Objective
As 2017 progresses, hyper-converged software will continue to grow in popularity but to cement its success, users will need to be able take full advantage of its productivity promise. The incredible power of parallel processing software will enable users to take advantage of what their hardware and software can do (see this video from ESG as an example).
Hyper-converged systems today are in essence a server plus a software-defined infrastructure, but often they are severely restricted in terms of performance and use cases and too often lack needed flexibility and a path for integration within the larger IT environment (for instance not supporting fibre channel, which often is key to enterprise and database connectivity). Powerful software-defined storage technologies that can do parallel I/O effectively provide a higher level of flexibility and leverage the power of multicore servers so fewer nodes are needed to get the work done, making them more cost-effective. Likewise, the software can incorporate existing flash and disk storage without creating additional silos; migrate and manage data across the entire storage infrastructure; and effectively utilize data stored in the cloud.
Data infrastructures including hyper-converged systems can all benefit from these advances through advanced parallel I/O software technologies that can dramatically increase their productivity by untapping the power that lies within standard multicore servers. While hyper-converged has become the buzzword of the day, let's remember the real objective is to achieve the most productivity at the lowest cost, therefore better utilization of one's storage and servers to drive applications is the key.
The Next Giant Leap Forward - Leveraging the Multiplier Impact of Parallel Processing on Productivity
This combination of powerful software and servers will drive greater functionality, more automation, and comprehensive services to productively manage and store data across the entire data infrastructure. It will lead to a new era where the benefits of multicore parallel processing can be applied universally. These advances (which are already before us) are key to solving the problems caused by slow I/O and inadequate response times that have been responsible for holding back application workload performance and cost savings from consolidation. The advances in multicore processing, parallel processing software and software-defined infrastructure, collectively, are fundamental to achieving the next giant leap forward in business productivity.
No comments:
Post a Comment