GPU Instance provides cost-effective, accessible high-performance computing with easy GPU access, making it ideal for complex AI tasks. Users benefit from on-demand rental, optimizing both efficiency and cost-effectiveness.

Features

  • Rich built-in templates help you quickly deploy GPU instances for AI tasks out of the box;
  • Start a GPU instance in seconds, helping you quickly handle traffic peaks;
  • Provides nodes in multiple global regions to deploy GPUs closer to your users for minimal latency;
  • Provides free local storage and allows you to attach network volume, also provides free network transfer;
  • Billing is accurate to the second, charging only for the actual running time of the GPU container instances;
  • Affordable pricing helps you save costs by up to 50% easily.

Terminology

  • Container Image: Docker-compatible OCI image reference, supporting both public registries and private repositories with authentication.
  • Environment Variables: Runtime configuration key-value pairs injected into containers for dynamic configuration and secrets management.
  • HTTP/TCP Ports: Network endpoints exposed by containers for communication. HTTP ports (e.g., 80, 443) for web services, TCP ports for general network protocols.
  • Container Disk: Root filesystem partition containing OS and system binaries. Mounted as read-write at container startup.
  • Volume Disk: High-performance ephemeral storage directly attached to the GPU instance. Optimized for I/O intensive workloads but non-persistent across instance restarts.
  • Network Volume: Network-attached storage service that allows users to access and manage data through a network. Data can be stored on remote servers and accessed from any device.

If you have any questions, you can first check the FAQs for GPU Instance.

Was this page helpful?