Clusters are ideal for interactive development, debugging, and building new projects. If you have working code and want to run experiments, Jobs are recommended instead.
Make sure you’ve installed the TensorPool CLI and configured your API key.
Create Your First GPU Cluster
Create a 1xB200For multi-node training, create a 4-node 8xB200 cluster:You can also create a cluster with a pre-built container image (includes CUDA, Python, and ML libraries):
See instance types for all available GPU configurations and container images for available images.
Check Your Cluster Status
The Wait until the status shows
tp cluster create command will give you a cluster ID (e.g., c-abc123). Use it to check your cluster’s status:RUNNING. The output will list your cluster’s instances, each with an instance ID (e.g., i-xyz789)If you lose the cluster ID, you can always find it with
tp cluster listSSH Into Your Cluster
Once your cluster status is For multi-node clusters, SSH into the jumphost instance first. From there, you can access worker nodes by name (e.g.,
RUNNING, grab the instance ID from tp cluster info and connect:ssh <cluster_id>-0).Next Steps
- Learn about cluster management
- Explore storage volumes for persistent data
- Check out multi-node training for distributed workloads