对象存储

Vultr 对象存储 – 面向 AI 工作负载的 S3 兼容存储

可扩展的 S3 兼容对象存储,与 Vultr GPU 实例集成。以实惠的价格存储训练数据集、模型检查点和生产工件。

Explore GPU Cloud →

Object Storage Features

🔌

S3-Compatible API

Drop-in replacement for AWS S3. Use boto3, rclone, s3cmd, MinIO client, or the AWS CLI with a single endpoint change — no code rewrite needed.

💰

Predictable Per-GB Pricing

Pay only for what you store. No per-request charges for standard GET/PUT operations. Ideal for large-scale ML dataset storage with high I/O frequency.

🌐

Global CDN Edge

Vultr's anycast CDN delivers static assets from 32+ global PoPs. Serve model inference responses, static ML artifacts, and API results at the edge.

🔒

Access Control & Encryption

Fine-grained S3 ACLs, bucket policies, and server-side AES-256 encryption. CORS configuration for web-facing API integrations.

GPU Instance Integration

Mount buckets via s3fs-fuse or stream data via boto3 directly from Vultr GPU instances. Co-located storage and compute minimize egress latency.

♾️

Unlimited Scalability

No object count limits. Store petabytes of training data, model weights, and checkpoints without pre-provisioning storage capacity.

Object Storage Pricing vs AWS S3 & GCS

FeatureVultrAWS S3Google Cloud
Storage (per GB/mo)~$0.020~$0.023 (S3 Standard)~$0.020 (Standard)
GET requests (10K)Free$0.004$0.004
PUT requests (1K)Free$0.005$0.005
Egress (per GB)~$0.01 (CDN)$0.09 (Internet)$0.08 (Internet)
S3 API CompatibleYesNativeVia XML API
GPU Co-locationYes (same DC)PartialPartial

AI & ML Storage Use Cases

🧠

ML Training Datasets

Store ImageNet, Common Crawl, or custom datasets. Stream multi-TB datasets directly to GPU training nodes using tf.data or PyTorch DataLoader with S3 connectors.

⚖️

Model Weight Storage

Centralize GGUF, safetensors, ONNX, and checkpoint files. Version model weights with bucket versioning and restore previous checkpoints instantly.

🎬

Generative AI Assets

Store raw video/audio for fine-tuning multimodal models. Serve generated images and videos via Vultr CDN to end users without bandwidth spikes.

💾

Database Backups

Automated PostgreSQL, MongoDB, and Redis backups to object storage. Lifecycle rules automatically archive old backups to cold storage tiers.

Quick Integration Examples

boto3 (Python)python
import boto3

s3 = boto3.client(
    "s3",
    endpoint_url="https://ewr1.vultrobjects.com",
    aws_access_key_id="YOUR_ACCESS_KEY",
    aws_secret_access_key="YOUR_SECRET_KEY",
)

# Upload dataset
s3.upload_file("dataset.tar.gz", "my-ml-bucket", "datasets/dataset.tar.gz")

# Stream to GPU instance
obj = s3.get_object(Bucket="my-ml-bucket", Key="models/llama3-70b.gguf")
data = obj["Body"].read()
rclone syncbash
# Configure rclone
rclone config create vultr s3 \
  provider=Other \
  endpoint=ewr1.vultrobjects.com \
  access_key_id=YOUR_KEY \
  secret_access_key=YOUR_SECRET

# Sync datasets to GPU instance
rclone sync vultr:my-ml-bucket/datasets ./datasets/ \
  --transfers=32 --checkers=16 --progress

Related Technical Guides

Related Infrastructure Pages

对象存储常见问题

Vultr 对象存储与 AWS S3 兼容吗?

是的。Vultr 对象存储使用 S3 兼容 API,这意味着任何与 AWS S3 配合使用的工具(boto3、rclone、s3fs、MinIO 客户端和 AWS CLI)只需简单更改端点即可原生与 Vultr 对象存储配合使用。

与 AWS S3 相比定价有何不同?

Vultr 对象存储比 AWS S3 便宜得多。Vultr 按 GB 存储收费,标准操作无按请求费用,非常适合频繁 GPU 实例访问的大型 ML 数据集存储。

我可以将 Vultr 对象存储直接挂载到 GPU 实例吗?

可以。您可以使用 s3fs-fuse 或 goofys 挂载 Vultr 对象存储,或在训练脚本中通过 Python (boto3) 使用 S3 兼容 API 访问它。这使得无需本地磁盘瓶颈即可流式传输大型数据集。

哪些数据类型最适合 Vultr 对象存储?

ML 训练数据集、模型权重文件(GGUF、safetensors)、推理工件、生成式 AI 的视频资产、数据库备份和静态网站资产都非常适合对象存储。

Ready to Store AI Datasets on Vultr?

New accounts signed up via referral link may be eligible for promotional credits. Credits subject to Vultr's official program terms.