Docs
DataCrunch HomeSDKAPILogin / Signup
  • Welcome to DataCrunch
    • Overview
    • Locations and Sustainability
    • Support
  • GPU Instances
    • Set up a GPU instance
    • Securing Your Instance
    • Shutdown, Hibernate, and Delete
    • Adding a New User
    • Block Volumes
    • Shared Filesystems (SFS)
    • Managing SSH Keys
    • Connecting to Your DataCrunch.io Server
    • Connecting to Jupyter notebook with VS Code
    • Team Projects
    • Pricing and Billing
  • Clusters
    • Instant Clusters
      • Deploying a GPU cluster
      • Slurm
      • Spack
      • Good to know
    • Customized GPU clusters
  • Containers
    • Overview
    • Container Registries
    • Scaling and health-checks
    • Batching and Streaming
    • Async Inference
    • Tutorials
      • Quick: Deploy with vLLM
      • In-Depth: Deploy with TGI
      • In-Depth: Deploy with SGLang
      • In-Depth: Deploy with vLLM
      • In-Depth: Deploy with Replicate Cog
      • In-Depth: Asynchronous Inference Requests with Whisper
  • Inference
    • Overview
    • Authorization
    • Audio Models
      • Whisper X
  • Pricing and Billing
  • Resources
    • Resources Overview
    • DataCrunch API
  • Python SDK
  • Get Free Compute Credits
Powered by GitBook
On this page
  • Deploying on-demand cluster
  • Accessing your cluster
  • Running jobs

Was this helpful?

  1. Clusters
  2. Instant Clusters

Deploying a GPU cluster

Last updated 25 days ago

Was this helpful?

Deploying on-demand cluster

In your , select Clusters->Deploy cluster. In the next screen you can select your contract duration (starting from 1 week) and the number of nodes (from two to eight) you would like to have in your cluster.

Next, select your shared filesystem size. File systems are mounted as follows:

  • Local storage is mounted to /mnt/local_disk on each worker node.

  • SFS is mounted to /home on all nodes, including the jump host.

You also need to supply your SSH public key before you deploy. We recommend you choose the cluster hostname appropriately, since your worker nodes will inherit the hostname as the prefix.

Accessing your cluster

Please also note that the jump host node will become accessible a few minutes before the worker nodes are ready, when starting the cluster for the first time.

The default Linux user for your on-demand cluster is ubuntu

Once the cluster has been created, you can proceed to log in by copying the ssh ubuntu@CLUSTER_IP command from the Clusters screen in your dashboard.

You can login to the individual worker nodes from your jumphost by running ssh WORKER_NAME

You can use tab-completion with SSH to quickly login to your worker nodes

Running jobs

Once the above steps are done, you deploy the cluster, just like you would an .

Once deployment has been done on , please give the cluster around 20 minutes to start.

We recommend using to run jobs on the cluster.

ordinary DataCrunch cloud instance
DataCrunch cloud dashboard
Slurm
DataCrunch cloud dashboard