Q&A: High-performance flash key to unlocking data-intense AI workloads – SiliconANGLE News

Artificial-intelligence workloads are intrinsically data-intensive. They need massive amounts of data to train and produce valuable insights. So storage becomes a key consideration for running these AI applications. First, they need to run where the data resides, which varies in distributed environments. And then those informationalgems that result from the AI algorithms need to be stored somewhere.

Flash storage has a multi-dimensional performance that can take any size of file or workload and run through it without creating storage-related bottlenecks. Two IT leaders in high-performance computing, Nvidia Corp.and Pure Storage Inc., saw these demands in their customers. In response, they joined forces and created AIRI, an AI-ready infrastructure that can help unlock data intelligence.

You know, a lot of it comes from our customers, saidCharlie Boyle(pictured right), vice president and general manager of DGX Systems at Nvidia. Thats how we first started with Pure. Its our joint customer saying we need this stuff to work really fast. Theyre making a massive investment with us in computing. And so if youre going to run those systems at 100%, you need storage that can feed them.If the customer has data, we want it to be as simple as possible for them to run AI.

Boyle and Brian Schwarz (pictured left), vice president of product management at Pure Storage, spoke with Dave Vellante (@dvellante) and Lisa Martin (@LisaMartinTV), co-hosts of theCUBE, SiliconANGLE Medias mobile livestreaming studio, during the Pure//Accelerate event in Austin, Texas. They discussed theNvidia and Pure Storage partnership, the adoption of AIRI, and AI advancements in the industry(see the full interview with transcript here). (* Disclosure below.)

[Editors note: The following answers have been condensed for clarity.]

Martin: Give us a little bit of an overview of where Pure and Nvidia are.

Schwarz: It really was born out of work with mutual customers. We brought out the FlashBlade product. Obviously, Nvidia was in the market with DGXs for AI, and we really started to see overlap in a bunch of initial [AI] deployments. So thats really kind of where the partnership was born. And, obviously, the AI data hub is the piece that we really talked about at this years Accelerate.

Martin: Tell us a little bit about the adoption [of AIRI] and what customers are able to do with this AI-ready infrastructure?

Boyle: [Early customers] had been using storage for years, and AI was kind of new to them, and they needed that recipe. So the early customer experiences turned into AIRI the solution. And the whole point of it is to simplify AI.

AI sounds kind of scary to a lot of folks, and the data scientists really just need to be productive. They dont care about infrastructure, but IT has to support this. So IT was very familiar with Pure Storage. They used them for years for high-performance data, and as they brought in the Nvidiacompute to work with that, having a solution that we both supported was super important to the IT practitioners.

Vellante: How do you see the landscape? Are you seeing pretty aggressive adoption or is it still early?

Boyle: So, every customer is at a different point. Theres definitely a lot of people that are still early, but weve seen a lot of production use cases. So depending on the industry, it really depends on where people are in the maturity curve. But really our message out to the enterprise is start now, whether youve got one data scientist or youve got some community data scientists, theres no reason to wait on AI.

Vellante: So, what are the key considerations for getting started?

Schwarz: So I think understanding the business value creation problem is a really important step. And many people go through an early stage of experimentation a prototyping stage before they go into a mass-production use case. Its a very classic IT adoption curve.

If you look forward over the next 15 to 20 years, theres a massive amount of AI coming, and it is a new form of computing the GPU-driven computing. And the whole point about AIRI is getting the ingredients right to have this new set of infrastructure have storage, network, compute, and the software stack.

Martin: For other customers in different industries, how do you help them even understand the AI pipeline?

Boyle: A lot of it is understanding your data, and thats where Pure and the [AI] data hub comes in. And then formulate a question like, what could I do if I knew this thing? Because thats all about AI and deep learning. Its coming up with insights that arent natural when you just stare at the data. How can the system understand what you want. And then what are the things that you didnt expect to find that AI is showing you about your data. AI can unlock things that you may not have pondered yourself.

And one of the biggest aha moments that Ive seen in customers in the past year or so is just how quickly by using GPU computing they can actually look at their data, do something useful with it, and then move on to the next thing. So, that rapid experimentation is what AI is all about.

Vellante: Youre not going to help but run into [machine intelligence]; its going to be part of your everyday life. Your thoughts?

Boyle: We all use AI every day; you just dont know it. Its the voice recognition system, getting your answer right the first time all in less than a second. Before thatd be like you talked to an IVR system, wait, then you go to an operator; and now people are getting such a better user experience out of AI-backed systems.

Vellante: The AI leaders are applying machine intelligence to that data. How has this modern storage that we heard about this morning affected that customers abilities to really put data at their core?

Schwarz: I think one of the real opportunities, particularly with flash, is to consolidate data into a smaller number of larger, kind of islands of data, because thats where you can really drive the insights. And historically in a disk-driven world, you would never try to consolidate your data, because there were too many bad performance implications of trying to do that. The difference with flash is theres so much performance at the core of it, at the foundation of it.

Martin: I want to ask you about distributed environments. You know customers have so much choice for everything these days. On-prem, hosted, SaaS, public cloud. What are some of the trends that youre seeing?

Schwarz: The first thing I always tell people is, wheres your data gravity? Moving very large sets of data is actually still a hard challenge today. So running your AI where your data is being generated is a good first principle. The second thing is about giving people flexibility. So trying to use a consistent set of infrastructure and software and tooling that allows people to migrate and change over time is an important strategy.

Vellante: So, ideally, on-prem versus cloud implementations shouldnt be different but are they today?

Boyle: So, at the lowest level, there are always technical differences, but at the layer that customers are using it, we run one software stack, no matter where youre running. Its the same in the Nvidiasoftware stack. And its really [about] running AI where your data is.

Vellante: Now that youve been in the market for a while, what are Pures competitive differentiators?

Schwarz: Why do we think [flash] is a great fit for an AI use case? One is the flexibility of the performance we call it multi-dimensional performance. Small files, large files, meta data-intensive workloads, FlashBladecan do them all. Its a ground-up design, its super flexible on performance, but also, more importantly, I would argue simplicity is a real hallmark of who we are.

Watch the complete video interview below, and be sure to check out more of SiliconANGLEs and theCUBEs coverage of the Pure//Accelerate event. (* Disclosure: TheCUBE is a paid media partner for the Pure//Accelerate event. Neither Pure Storage Inc., the sponsor for theCUBEs event coverage, nor other sponsors have editorial control over content on theCUBE or SiliconANGLE.)

Wed like to tell you about our mission and how you can help us fulfill it. SiliconANGLE Media Inc.s business model is based on the intrinsic value of the content, not advertising. Unlike many online publications, we dont have a paywall or run banner advertising, because we want to keep our journalism open, without influence or the need to chase traffic.The journalism, reporting and commentary onSiliconANGLE along with live, unscripted video from our Silicon Valley studio and globe-trotting video teams attheCUBE take a lot of hard work, time and money. Keeping the quality high requires the support of sponsors who are aligned with our vision of ad-free journalism content.

If you like the reporting, video interviews and other ad-free content here,please take a moment to check out a sample of the video content supported by our sponsors,tweet your support, and keep coming back toSiliconANGLE.

Continue reading here:

Q&A: High-performance flash key to unlocking data-intense AI workloads - SiliconANGLE News

Related Posts

Comments are closed.