Together

Together AI provides large-scale GPU clusters equipped with NVIDIA's latest Blackwell and Hopper architecture GPUs, offering services primarily aimed at AI training and inference workloads. The clusters are interconnected using NVLink and InfiniBand, and they utilize advanced storage solutions and orchestration through Kubernetes and Slurm to deliver specialized and optimized AI computing resources.

Together

Provider Profile

Founded

Undated

Headquarters

San Francisco, CA

Pricing Model

Reserved Capacity and On-demand Capacity providing flexibility in pricing with options for discounts and variable billing based on usage.

Technical Specification

Target Audience
  • Enterprise ML
  • Researchers
  • Organizations needing large-scale AI compute
GPU Clusters & Offerings
  • NVIDIA GB200, B200, H200, H100 GPUs
  • Together Kernel Collection for accelerated AI computations
  • NVIDIA GB200 NVL72 liquid-cooled supercomputer
  • Custom GPU clusters design
  • NVMe and high-performance converged storage
Network Fabric
  • InfiniBand
  • Ethernet
Connectivity Bandwidth
14.4 Tbps Infiniband
Storage Architecture
  • NVMe SSDs
  • High-performance converged storage
  • VAST Data
  • WEKA AI-native storage systems
Compute Framework Compatibility
PyTorch
Resource Orchestration
  • Kubernetes
  • Slurm
Developer Interface & APIs
  • CLI
  • SDK
  • REST API
  • Terraform provider
Support Operations
  • Kubernetes Dashboard access
  • Direct SSH access
  • Support contact options
Resource Availability
GA
Datacenter Locations
Key Platform Features
  • NVLink and InfiniBand networking
  • Together Kernel Collection for enhanced training and inference speeds
  • Expert AI advisory services
  • Custom-built NVIDIA Blackwell clusters
  • VAST Data and WEKA AI-native storage systems
  • AI Data Centers with significant power capacity

Last Audit: February 2026