Home / Stream/ Setting Up Cribl Stream/Deployment Planning

Deployment Planning

There are at least three key factors that will determine the type of Cribl Stream deployment in your environment:

  • Amount of Incoming Data: This is defined as the amount of data planned to be ingested per unit of time. E.g., how many MB/s or GB/day?

  • Amount of Data Processing: This is defined as the amount of processing that will happen on incoming data. E.g., are there a lot of transformations, regex extractions, parsing functions, field obfuscations, etc.?

  • Routing and/or Cloning: Is most data going to a single destination, or is it being cloned and routed to multiple places? This is important because destination-specific serialization tends to be relatively expensive.

These factors are covered in detail in Sizing and Scaling, and in our Architectural Considerations introduction to reference architectures.

Type of Deployment

OS and System Requirements

Leader and Worker Nodes should have sufficient CPU, RAM, network, and storage capacity to handle your specific workload. It’s very important to test this before deploying to production. For details, see OS and System Requirements.

Cluster Installation/Configuration Checklist

This section compiles basic checkpoints for successfully launching a distributed cluster.

1. Provision Hardware

2. Configure Leader Node

3. Configure Worker Nodes

4. Map Workers to Groups

  • On the Leader Node, create a Worker Group.
    • Name the Worker Group (arbitrarily) POV.
  • On the Leader Node, confirm that workers are connecting.
    • In the sidebar, select Workers.
  • Map Workers to dev Worker Groups.
    • Use the Filter: cribl.tags.includes('POV').

5. Other

If you will be using Cribl Stream’s GeoIP enrichment feature, install the MaxMind database onto the Cribl Stream Leader and all Worker Nodes.