Docs
DataCrunch HomeSDKAPILogin / Signup
  • Welcome to DataCrunch
    • Overview
    • Locations and Sustainability
    • Pricing and Billing
    • Team Projects
    • Support
  • CPU and GPU Instances
    • Set up a CPU or GPU instance
    • Securing Your Instance
    • Shutdown, Hibernate, and Delete
    • Adding a New User
    • Managing SSH Keys
    • Connecting to Your DataCrunch.io Server
    • Connecting to Jupyter notebook with VS Code
  • Clusters
    • Instant Clusters
      • Deploying a GPU cluster
      • Slurm
      • Spack
      • Good to know
    • Customized GPU clusters
  • Storage
    • Block Volumes
      • Attaching a block volume
      • Resizing a block volume
      • Cloning a block volume
      • Permanently deleting a block volume
    • Shared Filesystems (SFS)
      • Creating a shared filesystem
      • Editing share settings
      • Mounting a shared filesystem
  • Containers
    • Overview
    • Container Registries
    • Scaling and health-checks
    • Batching and Streaming
    • Async Inference
    • Tutorials
      • Quick: Deploy with vLLM
      • In-Depth: Deploy with TGI
      • In-Depth: Deploy with SGLang
      • In-Depth: Deploy with vLLM
      • In-Depth: Deploy with Replicate Cog
      • In-Depth: Asynchronous Inference Requests with Whisper
  • Inference
    • Overview
    • Authorization
    • Image Models
      • Flux.1 Kontext pro
      • Flux.1 Kontext max
    • Audio Models
      • Whisper X
  • Pricing and Billing
  • Resources
    • Resources Overview
    • Services Overview
    • Shared Responsibility Model
    • DataCrunch API
  • Python SDK
  • Get Free Compute Credits
Powered by GitBook
On this page
  • Bare Metal
  • Instant Clusters
  • GPU and CPU Instances
  • Storage
  • Serverless Containers
  • Managed endpoints

Was this helpful?

  1. Resources

Services Overview

Last updated 1 day ago

Was this helpful?

DataCrunch offers a variety of services for our customers. We offer on-demand access, as well as direct and long-term access, to high-end GPU compute, CPU compute, Storage, Serverless Containers, and Managed Inference. Below is an outline of each of our services and supporting materials.

Bare Metal

Bare metal contracts provide the customer with sole access to an entire physical server on a contract-only basis. These can be in an arrangement of clusters or singular full servers. For instance, a customer may book a 128-GPU H200 cluster for six months. This would be specifically prepared for the client based on their requirements and can include adjustments to some hardware specifications and additional support services.

These bare metal servers are non-virtual devices and are fully assigned to the customer. Bare Metal customers do not share their device with any other customers and are the sole tenant within the physical server. DataCrunch is responsible for maintaining the power, networking, and physical operation of the server.

Instant Clusters

are virtualized on-demand clusters. In this arrangement, DataCrunch is responsible for the power, networking, physical operation of the hardware, and software operation of the host machines, as well as the initial setup of the virtual machine instances comprising the cluster. However, software updates and internal configuration changes inside the virtual machine instances are the customer's responsibility.

GPU and CPU Instances

On-demand are available via the cloud dashboard. These are available in numerous configurations with multiple GPU models, numbers of GPUs, and CPU-core counts. These instances are available as on-demand (pay per usage) or as long-term deployments (one month to two years). On-demand instances are available as:

  1. Uninterruptible fixed or dynamically-priced instances,

  2. Interruptible spot instances, offered at a discounted rate.

Spot instances may be discontinued at any point and allocated to on-demand or other services. This may mean that a spot instance is removed even a few seconds after creation.

In contract agreements, it is possible to arrange a credit limit or outline a specific arrangement of instances instead of pre-paying for services, as is the case with our self-service agreements.

Storage

We offer two types of Storage solutions, and a . The Block Volumes are required for the operation of an instance, as these are the OS drives. The shared file system can be accessed by multiple instances simultaneously.

Serverless Containers

Customers are responsible only for the internal operation of their containers. They may send traffic to the endpoint through our systems or through their services hosted elsewhere.

Managed endpoints

are a managed service in which DataCrunch hosts and scales customers' containers. A unique endpoint with bearer-token authentication is created for each container that the customer hosts with DataCrunch. The customer is expected to build their own services that make requests to the container endpoints.

The service provides API access to different models via managed endpoints that are run by DataCrunch. Here, the customer is only responsible for the inputs they provide to such endpoints. Managed endpoints may at times rely on 3rd party APIs, which will be clearly stated in relation to each respective endpoint.

Serverless Containers
Block Volumes
Shared File System
Instant Clusters
GPU and CPU instances
Inference