How IT teams can set the right foundation for AI projects – Medium

Goal #1: Support a Range of Applications

An AI platform doesnt just need to support TensorFlow or even just the model development workloads. It needs to provide testing pipelines, versioning, sandbox environments, monitoring, and more.

For example, you might start creating Kubernetes clusters for AI workloads. That cluster will run a wide set of applications that need access to a variety of datasets and compute hardware and likely even a variety of protocols.

Like with any platform hosted by IT and DevOps teams, an AI platform should support application scalability and resiliency. And, optimally, data scientists should have self-serve access to new environments.

Without a cohesive plan to support the production pipeline as a unified project, individual application silos often become inefficient, unscalable, and fragile.

Step back and ask, How can we make this set of disparate workloads as easy to manage and to scale as possible?

If youre an IT leader, you have an incredible opportunity. The success of your companys AI-fueled ambitions requires you to enable developers in a new way.

Get in front of the productionalization crisis by making architectural choices that will centralize AI infrastructure consolidating people, process and technology.

On the storage side, use the same centralized storage underneath all of the applications in the platform. For example, Pure Storages FlashBlade is great at handling all different IO patterns and has performant access for both file and object workloads, which means its well suited for any of these components.

Likewise, NVIDIAs DGX A100 brings consolidation to the compute hardware. With DGX A100, NVIDIA consolidated what used to be three separate silos of legacy compute infrastructure, each sized and designed for supporting only one specific workload: training or inference or analytics. DGX A100 supports all of these workflows using just one universal system type.

Now you have just two building blocks to manage one for storage and one for compute. This infrastructure simplicity is what lowers the threshold to be able to get models into production; theres already a place where new workloads can run. With the AIRI reference architecture from Pure Storage and NVIDIA, you can now support the end to end AI lifecycle from development to deployment on one elastic infrastructure.

See the original post here:

How IT teams can set the right foundation for AI projects - Medium

Related Posts

Comments are closed.