In today's enterprises, there is an ever-growing demand for AI-enabled services, from image and speech recognition to natural language processing, visual search and personalized recommendations. At the same time, datasets are exploding in size, networks are growing more complex and latency requirements are becoming more demanding as many AI applications now require responses in real time.
All of this is an argument for moving AI inferencing to the network edge - where the data is captured, the insights are gained and the action takes place. This is a point underscored in a recent report from Moor Insights & Strategy, titled "Delivering the AI-Enabled Edge with
"Nowhere is data as meaningful as the point at which it is generated," the firm notes in its report. "In many cases, the value of data is highest when insights are generated, and appropriate actions taken nearly instantaneously - sometimes in just fractions of a second. The ability to generate insights and actions at the edge, coupled with the widespread availability of affordable, high-performance AI processing capabilities, has led to AI being the number one workload for edge deployments. Indeed, some of the most exciting applications enabled by edge computing depend upon it."
Here's a case in point.
So how do you get there? For many organizations, the answer lies in the NVIDIA inference platform. This software is designed to deliver the performance, efficiency and responsiveness critical to powering the next generation of AI products and services - in the cloud, in the data center, at the network's edge and in autonomous machines.
This inference platform is designed to unleash the full potential of NVIDIA GPUs inside Dell EMC PowerEdge servers, which support both open-source NVIDIA Triton Inference Server software and NVIDIA TensorRT. The platform includes an inference optimizer and runtime that delivers low latency and high throughput for inference applications. This inference software is also a part of the NVIDIA AI Enterprise software suite, which is optimized, certified and supported by NVIDIA to enable customers to run AI workloads on
To bring it all together, Dell EMC systems are NVIDIA-Certified so enterprises can confidently deploy solutions that securely and optimally run their accelerated workloads. The certification test suite is designed to exercise the performance and functionality of the configured server or hyperconverged infrastructure by running a set of software that represents a wide range of real-world applications.
This certification process includes deep learning training, AI inference, AI frameworks including NVIDIA Riva and NVIDIA Clara, data science including NVIDIA RAPIDS and Spark, intelligent video analytics, high-performance computing and CUDA functions and rendering. It also covers infrastructure performance acceleration, such as network and storage offload, security features and remote management capabilities.
This is a point that Moor Insights & Strategy makes in its recent report, commissioned by
"IT organizations that are ready to embrace the edge need both a strategy and a partner that has experience with AI and the edge," the firm notes. "
For a look at the infrastructure to drive inference workloads at the edge, visit the
.
(C) 2021 M2 COMMUNICATIONS, source