Book Image

Splunk Best Practices

Book Image

Splunk Best Practices

Overview of this book

This book will give you an edge over others through insights that will help you in day-to-day instances. When you're working with data from various sources in Splunk and performing analysis on this data, it can be a bit tricky. With this book, you will learn the best practices of working with Splunk. You'll learn about tools and techniques that will ease your life with Splunk, and will ultimately save you time. In some cases, it will adjust your thinking of what Splunk is, and what it can and cannot do. To start with, you'll get to know the best practices to get data into Splunk, analyze data, and package apps for distribution. Next, you'll discover the best practices in logging, operations, knowledge management, searching, and reporting. To finish off, we will teach you how to troubleshoot Splunk searches, as well as deployment, testing, and development with Splunk.
Table of Contents (16 chapters)

Even data distribution


In most cases, the data distribution of every Forwarder is handled automatically within any Forwarder; however in some instances it is necessary to tell Splunk to force appropriate data distribution across the indexing layer. This is especially necessary when dealing with extremely large events and multiple indexing peers.

Let's take a clustered environment as an example:

Let's say that two of our API's are pulling data from systems that have events in excess of 10,000 lines per event, and they are updating pulling this data at 60-second intervals.

While Splunk can certainly do this, and a Heavy Forwarder can handle the data throughput, there are a few settings that need to be in place to make sure not only that performance is optimized, but also that the data is being load-balanced properly, is not being truncated before the data can finish.

There are a few things to know when consuming very large datasets at fast intervals:

  • By default, Splunk will only take in 1000...