Together
Together AI provides large-scale GPU clusters equipped with NVIDIA's latest Blackwell and Hopper architecture GPUs, offering services primarily aimed at AI training and inference workloads. The clusters are interconnected using NVLink and InfiniBand, and they utilize advanced storage solutions and orchestration through Kubernetes and Slurm to deliver specialized and optimized AI computing resources.
Together
Provider Profile
Founded
Undated
Headquarters
San Francisco, CA
Pricing Model
Reserved Capacity and On-demand Capacity providing flexibility in pricing with options for discounts and variable billing based on usage.
Technical Specification
Target Audience
- Enterprise ML
- Researchers
- Organizations needing large-scale AI compute
GPU Clusters & Offerings
- NVIDIA GB200, B200, H200, H100 GPUs
- Together Kernel Collection for accelerated AI computations
- NVIDIA GB200 NVL72 liquid-cooled supercomputer
- Custom GPU clusters design
- NVMe and high-performance converged storage
Network Fabric
- InfiniBand
- Ethernet
Connectivity Bandwidth
14.4 Tbps Infiniband
Storage Architecture
- NVMe SSDs
- High-performance converged storage
- VAST Data
- WEKA AI-native storage systems
Compute Framework Compatibility
PyTorch
Resource Orchestration
- Kubernetes
- Slurm
Developer Interface & APIs
- CLI
- SDK
- REST API
- Terraform provider
Support Operations
- Kubernetes Dashboard access
- Direct SSH access
- Support contact options
Resource Availability
GA
Datacenter Locations
Key Platform Features
- NVLink and InfiniBand networking
- Together Kernel Collection for enhanced training and inference speeds
- Expert AI advisory services
- Custom-built NVIDIA Blackwell clusters
- VAST Data and WEKA AI-native storage systems
- AI Data Centers with significant power capacity
Last Audit: February 2026