Using Grepr To Reduce Logging Costs

Jad Naous
April 25, 2025

Application logging is an essential technique to both confirm normal operations and determine root cause when an incident occurs. With modern applications becoming increasingly more complex the volume of log data produced has increased exponentially. This presents a challenge to businesses, logging data is undoubtedly useful but it is also expensive to process and store. This is true whether the log processing and aggregation is managed in house or handled by one of the many SaaS providers. The greater the volume of logs, the bigger the cost.

Not All Log Data Is Important All The Time

When everything is ticking along smoothly, logging is used to confirm that transactions are being processed in a prompt and error free manner. The actual content of the log message is not so important, instead the number of log messages per time period is often used as a metric; for example web server access logs. However, when an incident occurs every log line is potentially very important and under these circumstances there is no such thing as too much logging. This is a difficult balancing act to achieve, having enough information in logging to diagnose problems but not having too much log data so that the cost is prohibitive. With some applications it is possible to dynamically change the logging level, switching to more detailed logging when an incident is detected but this is a bit like closing the stable door after the horse has bolted; the important piece of information you need was not captured in the logs when the incident was triggered.

Two-Tier Storage

Grepr sits between log sources and log vendors, forwarding all log data to a data lake and sending summarized data to the vendor, with backfilling to fill historical data as needed.

Grepr slips in like a shim between the log shippers and the backend storage and aggregation server. It uses machine learning and a rules engine to determine what to forward to the backend and what to keep in low cost storage.

The machine learning algorithm automatically detects similarity patterns between the log messages then instead of passing every message through it periodically sends a summary of similar messages. Along with the summary of similar messages Grepr also sends samples of every pattern, therefore any unique or very infrequent message patterns are passed through without aggregation. This ensures that no important detail is not immediately available in your chosen log tooling.

Grepr automatically finds patterns in real-time and summarizes noisy patterns

The rules engine allows for fine tuning of what stays in low cost storage and what passes through. This typically results in a 90% reduction in log volume sent to the backend which realises a significant saving on log processing costs.

Nothing is discarded, every message either gets passed straight through when it matches a rule or it gets summarised and retained to low cost storage.

Dynamic Detail

The messages retained in low cost storage are not locked away, they are always available when needed. Grepr provides both a web interface and an API to query these logs. However, engineers do not want to look in two places when troubleshooting an incident; they want to stay with the log tool they know and love. This is where the magic of Grepr’s dynamic backfill comes to the rescue. The log messages relevant to the incident are quickly backfilled into the log aggregation server; dynamic low level detail on demand at low cost.

Start Saving Now

Give Grepr a spin and see how easy it is to start saving 90% on your logging services cost with zero interruption to your existing workflows. Stop worrying about achieving that fine balance between logging visibility and cost. You can now log in more detail without worrying about blowing the budget. Feel secure that you have all the log detail you need to deal with any incident in the tool you are already familiar with using Grepr dynamic backfill on demand.

Share this post

More blog posts

All blog posts
Product

How Grepr Handles Trace Logs

Grepr helps companies lower observability costs while keeping engineers' workflows unchanged. A key feature is its ability to ensure full logs for a sampled set of traces, even with log aggregation and sampling in place. Users can specify how much of their data to sample and provide trace ID paths to identify relevant logs. Grepr's system uses intelligent sampling, backfilling, and a rule engine to collect and process trace logs efficiently. To optimize performance and memory, it evolved from a basic set-based approach to a scalable design using Bloom filters and a custom resizing Bloom filter. This allows Grepr to maintain high throughput while managing memory use effectively, ensuring reliable trace log visibility.
May 14, 2025
Product

Automating Log Management

Grepr uses machine learning to reduce log volume without losing visibility. It parses and structures logs in real time, groups similar messages, and applies smart sampling to cut noise. Critical logs still get through, and full raw data is stored separately for easy access during incidents—keeping your backend lean and your team in control.
May 9, 2025
Product

Time Travel With Dynamic Backfill

Grepr’s Dynamic Backfill feature lets teams retain all log data at low cost while only sending essential logs to their main logging backend, cutting processing and storage costs by around 90%. Unlike traditional logging that risks missing key data before an incident is detected, Grepr stores everything in affordable storage and allows engineers to selectively backfill detailed logs when issues arise—like turning up log detail after the fact. This ensures engineers have full context for debugging, with no disruption to existing workflows, balancing deep visibility with major cost savings.
May 2, 2025

Get started free and see Grepr in action in 20 minutes.