cuDNN
/ˌsiː-juː-diː-ɛn-ɛn/
n. “A GPU-accelerated library for deep neural networks developed by NVIDIA.”
cuDNN, short for CUDA Deep Neural Network library, is a GPU-accelerated library created by NVIDIA that provides highly optimized implementations of standard routines used in deep learning. It is designed to work with CUDA-enabled GPUs and is commonly integrated into frameworks such as TensorFlow, PyTorch, and MXNet to accelerate training and inference of neural networks.
TensorRT
/ˈtɛnsər-ɑːr-ti/
n. “A high-performance deep learning inference library for NVIDIA GPUs.”
TensorRT is a platform developed by NVIDIA that optimizes and accelerates the inference of neural networks on GPUs. Unlike training-focused frameworks, TensorRT is designed specifically for deploying pre-trained deep learning models efficiently, minimizing latency and maximizing throughput in production environments.
NVIDIA
/ɛnˈvɪdiə/
n. “An American technology company specializing in GPUs and AI computing platforms.”
NVIDIA is a leading technology company known primarily for designing graphics processing units (GPUs) for gaming, professional visualization, and data centers. Founded in 1993, NVIDIA has expanded its focus to include high-performance computing, artificial intelligence, deep learning, and autonomous vehicle technologies.