DigitalOcean Spaces
Synopsis
Creates a target that writes log messages to DigitalOcean Spaces with support for various file formats and authentication methods. The target handles large file uploads efficiently with configurable rotation based on size or event count. DigitalOcean Spaces provides simple, scalable object storage with built-in CDN integration.
Schema
- name: <string>
  description: <string>
  type: digitaloceans3
  pipelines: <pipeline[]>
  status: <boolean>
  properties:
    key: <string>
    secret: <string>
    region: <string>
    endpoint: <string>
    part_size: <numeric>
    bucket: <string>
    buckets:
      - bucket: <string>
        name: <string>
        format: <string>
        compression: <string>
        extension: <string>
        schema: <string>
    name: <string>
    format: <string>
    compression: <string>
    extension: <string>
    schema: <string>
    max_size: <numeric>
    batch_size: <numeric>
    timeout: <numeric>
    field_format: <string>
    interval: <string|numeric>
    cron: <string>
    debug:
      status: <boolean>
      dont_send_logs: <boolean>
Configuration
The following fields are used to define the target:
| Field | Required | Default | Description | 
|---|---|---|---|
| name | Y | Target name | |
| description | N | - | Optional description | 
| type | Y | Must be digitaloceans3 | |
| pipelines | N | - | Optional post-processor pipelines | 
| status | N | true | Enable/disable the target | 
DigitalOcean Spaces Credentials
| Field | Required | Default | Description | 
|---|---|---|---|
| key | Y | - | DigitalOcean Spaces access key | 
| secret | Y | - | DigitalOcean Spaces secret key | 
| region | Y | - | DigitalOcean region (e.g., nyc3,sfo3,ams3,sgp1) | 
| endpoint | Y | - | Spaces endpoint URL (format: https://<region>.digitaloceanspaces.com) | 
Connection
| Field | Required | Default | Description | 
|---|---|---|---|
| part_size | N | 5 | Multipart upload part size in megabytes (minimum 5MB) | 
| timeout | N | 30 | Connection timeout in seconds | 
| field_format | N | - | Data normalization format. See applicable Normalization section | 
Files
| Field | Required | Default | Description | 
|---|---|---|---|
| bucket | N* | - | Default Spaces bucket name (used if bucketsnot specified) | 
| buckets | N* | - | Array of bucket configurations for file distribution | 
| buckets.bucket | Y | - | Spaces bucket name | 
| buckets.name | Y | - | File name template | 
| buckets.format | N | "json" | Output format: json,multijson,avro,parquet | 
| buckets.compression | N | "zstd" | Compression algorithm | 
| buckets.extension | N | Matches format | File extension override | 
| buckets.schema | N* | - | Schema definition file path (required for Avro and Parquet formats) | 
| name | N | "vmetric.{{.Timestamp}}.{{.Extension}}" | Default file name template when bucketsnot used | 
| format | N | "json" | Default output format when bucketsnot used | 
| compression | N | "zstd" | Default compression when bucketsnot used | 
| extension | N | Matches format | Default file extension when bucketsnot used | 
| schema | N | - | Default schema path when bucketsnot used | 
| max_size | N | 0 | Maximum file size in bytes before rotation | 
| batch_size | N | 100000 | Maximum number of messages per file | 
* = Either bucket or buckets must be specified. When using buckets, schema is conditionally required for Avro and Parquet formats.
When max_size is reached, the current file is uploaded to Spaces and a new file is created. For unlimited file size, set the field to 0.
Scheduler
| Field | Required | Default | Description | 
|---|---|---|---|
| interval | N | realtime | Execution frequency. See Interval for details | 
| cron | N | - | Cron expression for scheduled execution. See Cron for details | 
Debug Options
| Field | Required | Default | Description | 
|---|---|---|---|
| debug.status | N | false | Enable debug logging | 
| debug.dont_send_logs | N | false | Process logs but don't send to target (testing) | 
Details
The DigitalOcean Spaces target provides simple, developer-friendly object storage with comprehensive file format support. Spaces includes built-in CDN functionality powered by DigitalOcean's global network for fast content delivery.
Authentication
Requires DigitalOcean Spaces access credentials. Access keys can be generated through the DigitalOcean Control Panel under API settings. Each key provides full access to all Spaces in the account.
Endpoint Configuration
The endpoint URL follows the pattern https://<region>.digitaloceanspaces.com where <region> is your chosen DigitalOcean datacenter location.
Available Regions
DigitalOcean Spaces is available in the following regions:
| Region Code | Location | 
|---|---|
| nyc3 | New York City, USA | 
| sfo3 | San Francisco, USA | 
| ams3 | Amsterdam, Netherlands | 
| sgp1 | Singapore | 
| fra1 | Frankfurt, Germany | 
| syd1 | Sydney, Australia | 
File Formats
| Format | Description | 
|---|---|
| json | Each log entry is written as a separate JSON line (JSONL format) | 
| multijson | All log entries are written as a single JSON array | 
| avro | Apache Avro format with schema | 
| parquet | Apache Parquet columnar format with schema | 
Compression
All formats support optional compression to reduce storage costs and transfer times. Compression is applied before upload.
| Format | Compression Options | 
|---|---|
| JSON/MultiJSON | zstd(default),gzip | 
| Avro | null,deflate,snappy,zstd | 
| Parquet | uncompressed,gzip,snappy,zstd,brotli,lz4 | 
File Management
Files are rotated based on size (max_size parameter) or event count (batch_size parameter), whichever limit is reached first. Template variables in file names enable dynamic file naming for time-based partitioning.
Templates
The following template variables can be used in file names:
| Variable | Description | Example | 
|---|---|---|
| {{.Year}} | Current year | 2024 | 
| {{.Month}} | Current month | 01 | 
| {{.Day}} | Current day | 15 | 
| {{.Timestamp}} | Current timestamp in nanoseconds | 1703688533123456789 | 
| {{.Format}} | File format | json | 
| {{.Extension}} | File extension | json | 
| {{.Compression}} | Compression type | zstd | 
| {{.TargetName}} | Target name | my_logs | 
| {{.TargetType}} | Target type | digitaloceans3 | 
| {{.Table}} | Bucket name | logs | 
Multipart Upload
Large files automatically use multipart upload protocol with configurable part size (part_size parameter). Default 5MB part size balances upload efficiency and memory usage.
Multiple Buckets
Single target can write to multiple Spaces buckets with different configurations, enabling data distribution strategies (e.g., raw data to one bucket, processed data to another).
Schema Requirements
Avro and Parquet formats require schema definition files. Schema files must be accessible at the path specified in the schema parameter during target initialization.
CDN Integration
DigitalOcean Spaces includes built-in CDN functionality. Files stored in Spaces can be served globally through the CDN with no additional configuration required.
Cost Structure
DigitalOcean Spaces offers predictable pricing with no egress fees within reasonable limits, making it cost-effective for applications with moderate data transfer requirements.
Examples
Basic Configuration
The minimum configuration for a JSON Spaces target:
targets:
  - name: basic_spaces
    type: digitaloceans3
    properties:
      key: "DO00ABC123XYZ456"
      secret: "abcdef0123456789abcdef0123456789abcdef0123456789abcdef0123456789"
      region: "nyc3"
      endpoint: "https://nyc3.digitaloceanspaces.com"
      bucket: "datastream-logs"
Multiple Buckets
Configuration for distributing data across multiple Spaces buckets with different formats:
targets:
  - name: multi_bucket_export
    type: digitaloceans3
    properties:
      key: "DO00ABC123XYZ456"
      secret: "abcdef0123456789abcdef0123456789abcdef0123456789abcdef0123456789"
      region: "sfo3"
      endpoint: "https://sfo3.digitaloceanspaces.com"
      buckets:
        - bucket: "raw-data-archive"
          name: "raw-{{.Year}}-{{.Month}}-{{.Day}}.json"
          format: "multijson"
          compression: "gzip"
        - bucket: "analytics-data"
          name: "analytics-{{.Year}}/{{.Month}}/{{.Day}}/data_{{.Timestamp}}.parquet"
          format: "parquet"
          schema: "<schema definition>"
          compression: "snappy"
Parquet Format
Configuration for daily partitioned Parquet files:
targets:
  - name: parquet_analytics
    type: digitaloceans3
    properties:
      key: "DO00ABC123XYZ456"
      secret: "abcdef0123456789abcdef0123456789abcdef0123456789abcdef0123456789"
      region: "ams3"
      endpoint: "https://ams3.digitaloceanspaces.com"
      bucket: "analytics-lake"
      name: "events/year={{.Year}}/month={{.Month}}/day={{.Day}}/part-{{.Timestamp}}.parquet"
      format: "parquet"
      schema: "<schema definition>"
      compression: "snappy"
      max_size: 536870912
High Reliability
Configuration with enhanced settings:
targets:
  - name: reliable_spaces
    type: digitaloceans3
    pipelines:
      - checkpoint
    properties:
      key: "DO00ABC123XYZ456"
      secret: "abcdef0123456789abcdef0123456789abcdef0123456789abcdef0123456789"
      region: "sgp1"
      endpoint: "https://sgp1.digitaloceanspaces.com"
      bucket: "critical-logs"
      name: "logs-{{.Timestamp}}.json"
      format: "json"
      timeout: 60
      part_size: 10
With Field Normalization
Using field normalization for standard format:
targets:
  - name: normalized_spaces
    type: digitaloceans3
    properties:
      key: "DO00ABC123XYZ456"
      secret: "abcdef0123456789abcdef0123456789abcdef0123456789abcdef0123456789"
      region: "fra1"
      endpoint: "https://fra1.digitaloceanspaces.com"
      bucket: "normalized-logs"
      name: "logs-{{.Timestamp}}.json"
      format: "json"
      field_format: "cim"
Debug Configuration
Configuration with debugging enabled:
targets:
  - name: debug_spaces
    type: digitaloceans3
    properties:
      key: "DO00ABC123XYZ456"
      secret: "abcdef0123456789abcdef0123456789abcdef0123456789abcdef0123456789"
      region: "nyc3"
      endpoint: "https://nyc3.digitaloceanspaces.com"
      bucket: "test-logs"
      name: "test-{{.Timestamp}}.json"
      format: "json"
      debug:
        status: true
        dont_send_logs: true