AI Buyer Insights:

Moog, an UKG AutoTime customer evaluated Workday Time and Attendance

Westpac NZ, an Infosys Finacle customer evaluated nCino Bank OS

Wayfair, a Korber HighJump WMS customer just evaluated Manhattan WMS

Swedbank, a Temenos T24 customer evaluated Oracle Flexcube

Citigroup, a VestmarkONE customer evaluated BlackRock Aladdin Wealth

Cantor Fitzgerald, a Kyriba Treasury customer evaluated GTreasury

Michelin, an e2open customer evaluated Oracle Transportation Management

Moog, an UKG AutoTime customer evaluated Workday Time and Attendance

Westpac NZ, an Infosys Finacle customer evaluated nCino Bank OS

Wayfair, a Korber HighJump WMS customer just evaluated Manhattan WMS

Swedbank, a Temenos T24 customer evaluated Oracle Flexcube

Citigroup, a VestmarkONE customer evaluated BlackRock Aladdin Wealth

Cantor Fitzgerald, a Kyriba Treasury customer evaluated GTreasury

Michelin, an e2open customer evaluated Oracle Transportation Management

List of Google Compute Engine A3 Customers

Apply Filters For Customers

Logo Customer Industry Empl. Revenue Country Vendor Application Category When SI Insight Insight Source
Character AI Professional Services 90 $35M United States Google Google Compute Engine A3 ML and Data Science Platforms 2023 n/a In 2023, Character AI deployed Google Compute Engine A3 as part of a Google Cloud partnership to scale its conversational AI platform within its ML and Data Science Platforms footprint. The implementation uses Google Compute Engine A3 VMs with NVIDIA H100 accelerators alongside Google Cloud TPUs to support model development and production serving for large language models, explicitly targeting ML/model training and inference workloads. The architecture combines Google Compute Engine A3 instances and TPU capacity into distributed training and inference clusters, with standard ML components such as training pipelines, model parallelism and sharded checkpointing, and inference serving layers. Google Compute Engine A3 is used for GPU-accelerated training and inference, while TPUs contribute to high throughput training runs, creating a hybrid accelerator fabric to balance development iteration and production latency needs. The deployment is implemented on Google Cloud infrastructure in the United States and is integrated into Character AI s conversational platform to enable rapid scale to millions of users. Operational scope focuses on ML/model training and inference workloads across the platform rather than specific departmental modules, with cloud-native orchestration and autoscaling patterns implied by the use of A3 VMs and TPUs. Governance centered on production serving clusters and controlled model rollouts to support new feature launches, aligning release workflows with accelerator provisioning and capacity planning. The partner announcement states the configuration improved training and inference efficiency and expanded capacity for new features, enabling faster development cycles and larger scale serving without specifying quantitative metrics.
Contextual AI Professional Services 50 $5M United States Google Google Compute Engine A3 ML and Data Science Platforms 2023 n/a In 2023, Contextual AI began using Google Compute Engine A3 as part of its ML and Data Science Platforms strategy with Google Cloud named as the preferred cloud provider. The announcement specifies use of GPU VMs including Google Compute Engine A3 (H100) and A2 (A100) for building and training enterprise large language models, with a focus on retrieval augmented generation enabled enterprise AI use cases in the United States and targeting ML model training and inference for enterprise applications. Contextual AI configures Google Compute Engine A3 for scalable model development workflows and production inference serving, implementing standard ML platform capabilities such as distributed GPU training clusters, data preprocessing and feature pipelines, fine tuning of LLMs, and retrieval augmented generation pipelines. The deployment emphasizes autoscaling compute to support iterative model development and deployment, and uses orchestration of training jobs to manage GPU resource utilization across development and production workloads. The implementation is centered on Google Cloud infrastructure and aligns inference endpoints with enterprise data retrieval layers to support RAG workflows without naming external vendors. Operational scope is the United States and the initiative is positioned to support enterprise application use cases by centralizing model training and inference on Google Compute Engine A3 while instituting controlled rollout and orchestration practices to govern model promotion into production.
Tabnine Professional Services 100 $10M Israel Google Google Compute Engine A3 ML and Data Science Platforms 2025 n/a In 2025, Tabnine deployed Google Compute Engine A3 to run real time inference for its AI assisted code completion product. The deployment uses Google Compute Engine A3 A3 High machine types with NVIDIA H100 GPUs to support production inference workloads that serve developer workflows in Israel and globally. Tabnine configured Google Kubernetes Engine clusters that include smaller A3 High machine types, containerized model servers, and GPU provisioning to host low latency model endpoints. The implementation leverages node pools sized for H100 GPU capacity and standard inference serving patterns common to ML and Data Science Platforms, including automated scaling of replica counts to respond to developer request traffic. Integrations center on Google Kubernetes Engine and Google Cloud GPUs as the primary compute and orchestration stack, with model serving and container orchestration aligned to cloud native operational practices. Operational coverage spans engineering and product teams responsible for runtime model management and developer experience globally and within Israel. Governance focused on configuration controls for node pool sizing and model placement to manage latency tradeoffs, and operational runbooks for GPU instance lifecycle and scaling. Tabnine reports that moving inference to the smaller A3 H100 configurations delivered an approximate 36% reduction in real time code assist model latency, an explicit ML performance outcome tied to the Google Compute Engine A3 deployment in its ML and Data Science Platforms architecture.
Showing 1 to 3 of 3 entries

Buyer Intent: Companies Evaluating Google Compute Engine A3

ARTW Buyer Intent uncovers actionable customer signals, identifying software buyers actively evaluating Google Compute Engine A3. Gain ongoing access to real-time prospects and uncover hidden opportunities.

Discover Software Buyers actively Evaluating Enterprise Applications

Logo Company Industry Employees Revenue Country Evaluated
No data found