Docs
DataCrunch HomeSDKAPILogin / Signup
  • Welcome to DataCrunch
    • Overview
    • Locations and Sustainability
    • Pricing and Billing
    • Team Projects
    • Support
  • CPU and GPU Instances
    • Set up a CPU or GPU instance
    • Securing Your Instance
    • Shutdown, Hibernate, and Delete
    • Adding a New User
    • Managing SSH Keys
    • Connecting to Your DataCrunch.io Server
    • Connecting to Jupyter notebook with VS Code
    • Remote desktop access
  • Clusters
    • Instant Clusters
      • Deploying a GPU cluster
      • Slurm
      • Spack
      • Good to know
    • Customized GPU clusters
  • Storage
    • Block Volumes
      • Attaching a block volume
      • Resizing a block volume
      • Cloning a block volume
      • Permanently deleting a block volume
    • Shared Filesystems (SFS)
      • Creating a shared filesystem
      • Editing share settings
      • Mounting a shared filesystem
  • Containers
    • Overview
    • Container Registries
    • Scaling and health-checks
    • Batching and Streaming
    • Async Inference
    • Tutorials
      • Quick: Deploy with vLLM
      • In-Depth: Deploy with TGI
      • In-Depth: Deploy with SGLang
      • In-Depth: Deploy with vLLM
      • In-Depth: Deploy with Replicate Cog
      • In-Depth: Asynchronous Inference Requests with Whisper
  • Inference
    • Overview
    • Authorization
    • Image Models
      • Flux.1 Kontext pro
      • Flux.1 Kontext max
    • Audio Models
      • Whisper X
  • Pricing and Billing
  • Resources
    • Resources Overview
    • Services Overview
    • Shared Responsibility Model
    • DataCrunch API
  • Python SDK
  • Get Free Compute Credits
Powered by GitBook
On this page
  • Deploying on-demand cluster
  • Accessing your cluster
  • Running jobs

Was this helpful?

  1. Clusters
  2. Instant Clusters

Deploying a GPU cluster

Last updated 2 months ago

Was this helpful?

Deploying on-demand cluster

In your DataCrunch cloud dashboard, select Clusters->Deploy cluster. In the next screen you can select your contract duration (starting from 1 week) and the number of nodes (from two to eight) you would like to have in your cluster.

Next, select your shared filesystem size. File systems are mounted as follows:

  • Local storage is mounted to /mnt/local_disk on each worker node.

  • SFS is mounted to /home on all nodes, including the jump host.

You also need to supply your SSH public key before you deploy. We recommend you choose the cluster hostname appropriately, since your worker nodes will inherit the hostname as the prefix.

Once the above steps are done, you deploy the cluster, just like you would an ordinary DataCrunch cloud instance.

Accessing your cluster

Once deployment has been done on DataCrunch cloud dashboard, please give the cluster around 20 minutes to start.

Please also note that the jump host node will become accessible a few minutes before the worker nodes are ready, when starting the cluster for the first time.

The default Linux user for your on-demand cluster is ubuntu

Once the cluster has been created, you can proceed to log in by copying the ssh ubuntu@CLUSTER_IP command from the Clusters screen in your dashboard.

You can login to the individual worker nodes from your jumphost by running ssh WORKER_NAME

You can use tab-completion with SSH to quickly login to your worker nodes

Running jobs

We recommend using Slurm to run jobs on the cluster.