Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.sawmills.ai/llms.txt

Use this file to discover all available pages before exploring further.

Supported Data Types

📘 Logs | 📈 Metrics | 🚦 Traces

Configuration

FieldTypeDefaultRequiredDescription
NameStringnonetrueUnique identifier within Sawmills.
RegionString”us-east-1”trueAWS region.
S3 BucketStringnonetrueS3 bucket name.
Role ARNStringnonefalseThe Role ARN to be assumed.
File PrefixStringnonefalsePrefix for the S3 key (root directory inside the bucket).
Output FormatStringOTLP JSONfalseFormat used to produce output data (see Output Format options below).

Advanced Options

FieldTypeDefaultRequiredDescription
REST API EndpointStringNonefalseOverrides the endpoint, instead of constructing it from region and s3_bucket.
S3 Force Path StyleBooleanfalsefalseSet this to true to force the request to use path-style addressing.
Use SSLBooleantruefalseSet this to false to disable SSL when sending requests.
CompressionStringnonefalseShould the file be compressed.

Output Format

The Output Format determines how data is serialized to AWS S3.
  • OTLP JSON (default): The OpenTelemetry Protocol format represented as JSON.
  • OTLP (Protobuf): The OpenTelemetry Protocol format represented as Protocol Buffers. A single protobuf message is written into each object.
  • Sumo Logic (JSON): The Sumo Logic Installed Collector Archive format (logs only).
  • Raw payload: Exports the log body as a string (logs only).
  • NDJSON (.json.gz): Newline-delimited JSON, gzipped, hourly-partitioned (dt=YYYYMMDD/hour=HH/…). One log record per line. Compatible with Datadog Logs Rehydration, Splunk Generic S3 input, AWS Glue / Splunk FSS3, and other archive readers. (Logs only.)
  • Parquet: Apache Parquet files, hourly-partitioned. Compatible with Splunk FSS3, Snowflake, AWS Athena, and other analytics tools. (Logs only.)

Snowflake Table Schema for Parquet

When using the Parquet output format, you can create a Snowflake table with the following schema:
CREATE TABLE PARQUET_LOGS (
    Date TIMESTAMP_LTZ,
    Service VARCHAR,
    Source VARCHAR,
    Host VARCHAR,
    Status VARCHAR,
    Message VARCHAR,
    Attributes OBJECT,
    Tags OBJECT
);

AWS Credential Configuration

Sawmills collector is running on a Kubernetes cluster and uses helm charts. To provide AWS credentials, pass extraEnvs in values.yaml as shown below:
extraEnvs:
  - name: AWS_ACCESS_KEY_ID
    value: "<YOUR AWS ACCESS KEY>"
  - name: AWS_SECRET_ACCESS_KEY
    value: "<YOUR AWS SECRET ACCESS KEY>"