AI Buyer Insights:

Michelin, an e2open customer evaluated Oracle Transportation Management

Westpac NZ, an Infosys Finacle customer evaluated nCino Bank OS

Citigroup, a VestmarkONE customer evaluated BlackRock Aladdin Wealth

Swedbank, a Temenos T24 customer evaluated Oracle Flexcube

Moog, an UKG AutoTime customer evaluated Workday Time and Attendance

Wayfair, a Korber HighJump WMS customer just evaluated Manhattan WMS

Cantor Fitzgerald, a Kyriba Treasury customer evaluated GTreasury

Michelin, an e2open customer evaluated Oracle Transportation Management

Westpac NZ, an Infosys Finacle customer evaluated nCino Bank OS

Citigroup, a VestmarkONE customer evaluated BlackRock Aladdin Wealth

Swedbank, a Temenos T24 customer evaluated Oracle Flexcube

Moog, an UKG AutoTime customer evaluated Workday Time and Attendance

Wayfair, a Korber HighJump WMS customer just evaluated Manhattan WMS

Cantor Fitzgerald, a Kyriba Treasury customer evaluated GTreasury

List of NVIDIA TensorRT Customers

Apply Filters For Customers

Logo Customer Industry Empl. Revenue Country Vendor Application Category When SI Insight Insight Source
Amazon Retail 1578000 $638.0B United States NVIDIA NVIDIA TensorRT Apps Development 2024 n/a In 2024 Amazon implemented NVIDIA TensorRT alongside NVIDIA Triton to accelerate a T5 NLP model used for real time spell correction in product search. This deployment is recorded under Apps Development and was executed on AWS in the United States to provide low latency, high throughput inference for a customer facing search service. The implementation integrated NVIDIA TensorRT as the inference optimization engine with NVIDIA Triton Inference Server as the serving layer, hosting T5 model variants and managing model versioning and GPU resource allocation. Configuration focused on GPU accelerated inference optimizations typical of TensorRT, including reduced precision and operator fusion to increase throughput and reduce per request compute, and operationalized the serving pipeline to handle synchronous inference requests from the search stack. Integration scope covered the retail product search workflow and the search engineering and customer experience functions, with the inference service deployed in AWS regions within the United States. According to NVIDIA's case study the configuration delivered a reported 5x inference speedup and achieved sub 50ms latency for the targeted inference path, outcomes cited by NVIDIA as the result of the Triton and NVIDIA TensorRT based acceleration.
Samsung Medison South Korea Life Sciences 1040 $439M South Korea NVIDIA NVIDIA TensorRT Apps Development 2021 n/a In 2021, Samsung Medison South Korea integrated NVIDIA TensorRT into the Intelligent Assist features of its V8 high end ultrasound systems under an Apps Development initiative to accelerate on device inference. NVIDIA TensorRT was embedded in device software to support medical imaging workflows, with the stated objective of improving medical image quality and clinician support for reading and diagnosis, and the integration was announced as a use case in South Korea aimed to reduce scan time and streamline workflows. The implementation was delivered as device software integration within the V8 ultrasound product line, leveraging NVIDIA TensorRT for inference optimization and low latency runtime for deep learning models used by Intelligent Assist. Operational coverage focused on medical imaging and clinician reading workflows on the V8 systems in South Korea, and the work emphasized embedding inference acceleration into existing device software stacks as described in NVIDIA's announcement.
Snap Media 5367 $4.6B United States NVIDIA NVIDIA TensorRT Apps Development 2021 n/a In 2021, Snap implemented NVIDIA TensorRT to accelerate machine learning inference for ad and content ranking. The deployment targeted advertising and monetization workflows in the United States and leveraged NVIDIA GPUs to run production ranking models at scale. NVIDIA TensorRT was used to optimize model runtime and inference pipelines within Snap's Apps Development stack, focusing on model graph optimization, precision reduction to FP16 and INT8, operator fusion, and efficient batching to reduce per request compute. These category aligned capabilities enabled Snap to deploy heavier, more accurate ranking models while constraining inference resource use. Integration occurred at the GPU inference server layer, embedding NVIDIA TensorRT into Snap's serving infrastructure and model orchestration flows for ad and content ranking. According to the vendor announcement Snap realized about a 50 percent improvement in inference cost efficiency and halved serving latency, outcomes that supported expanded model complexity for advertising monetization in the United States.
Professional Services 73 $8M United States NVIDIA NVIDIA TensorRT Apps Development 2021 n/a
Showing 1 to 4 of 4 entries

Buyer Intent: Companies Evaluating NVIDIA TensorRT

ARTW Buyer Intent uncovers actionable customer signals, identifying software buyers actively evaluating NVIDIA TensorRT. Gain ongoing access to real-time prospects and uncover hidden opportunities. Companies Actively Evaluating NVIDIA TensorRT for Apps Development include:

  1. Dawning Information Industry, a China based Manufacturing organization with 5991 Employees
  2. The Aerospace Corporation, a United States based Aerospace and Defense company with 4600 Employees
  3. NVIDIA, a United States based Manufacturing organization with 36000 Employees

Discover Software Buyers actively Evaluating Enterprise Applications

Logo Company Industry Employees Revenue Country Evaluated
No data found