Goldsky: 96% Reduction in Datadog Logging Costs

Jad Naous
June 6, 2025
Grepr dashboard showing Goldsky's pipeline overview with 97% total reduction across all pipelines, 1.87 billion events processed, and 67.8 million events forwarded. The interface displays a table of individual pipelines with their respective input volumes, output volumes, and reduction percentages.

Goldsky is Web3's realtime data platform. They help developers build dApps faster through high-performance blockchain indexing, instant subgraphs, and custom data streaming pipelines.

About six months ago, their team came to us with a familiar problem: log volumes had gotten out of hand, and the costs no longer matched the value. They were collecting and storing far more than they actually needed. We deployed Grepr shortly after that initial conversation, and within weeks, their Datadog logging bills dropped by 96%.

Paymahn Moghadasian, Lead Engineer at Goldsky, handled the deployment.

How the Rollout Worked

Goldsky manages their infrastructure with Terraform across separate staging and production environments, including their Datadog agents.

Paymahn started in staging. He created a pipeline and pointed the Datadog agents to Grepr in about 20 minutes. Volume there was light, around 8 messages per second, but even at that scale, Grepr achieved roughly 80% reduction.

He let it run for a week to build confidence before moving to production.

Production required more care. Rather than cutting over all at once, Paymahn used Datadog's dual-shipping capability. This let him add Grepr as a destination while still sending logs directly to Datadog, so nothing was at risk during the transition.

Here's the approach he took:

  1. Enable dual shipping in Datadog for logs
  2. For each service, add a filter in Grepr to drop all logs except the one being migrated
  3. Once that service's logs flow through Grepr correctly and show up in Datadog, add a Drop Rule to stop logs for that service that aren't coming from Grepr
  4. Tune the setup with exceptions as needed to preserve existing alerts and dashboards
  5. Run for a day to validate
  6. Move to the next service
  7. Optionally update alerts or dashboards to take advantage of summarized data instead of raw data
  8. After two weeks of validation, turn off dual-shipping from the agents

The full process took four weeks from start to finish.

The Numbers

For May 2025:

Indexed Logs: 5.7 billion messages dropped to 250 million, a 96% reduction

Ingested Logs: 12 terabytes dropped to 795 gigabytes, a 93% reduction

The dollar savings matched. After accounting for Grepr's costs, Goldsky cut their overall Datadog logging spend by more than 85%.

What About Troubleshooting?

When we asked Paymahn about impact on mean time to resolution, he gave us two words: "no impact."

His team actually found that with the noise filtered out, logs became easier to read and understand than before.

Other Wins

Time back for building: By solving the log cost problem quickly, Goldsky freed up engineering time for their actual product.

Historical search without rehydration: They could search logs across multiple months without paying extra to rehydrate archived data.

Cleaner signal: Less noise meant faster comprehension when reading through logs.

In Paymahn's Words

"Grepr's immediate, high-touch support was excellent. We always felt taken care of."

"The UI works well for what we need. It's not trying to compete with Datadog's UI, and that's fine."

"Grepr was always up and available."

"Logs arrived at Datadog with some minimal added latency, but nothing that mattered in any real way."

And his final take:

"Grepr allowed us to keep all our established observability use cases and processes intact by essentially getting rid of the noise in the data. Lower costs without any retraining."

Share this post

More blog posts

All blog posts
Grepr team members John and Utkarsh at an outdoor café in Amsterdam during KubeCon EU 2026.
Events

KubeCon Amsterdam 2026: Hallway Conversations Said What Keynotes Didn't

KubeCon EU 2026 made one thing clear: AI infrastructure is generating telemetry volumes that most observability budgets were never built to handle.
April 8, 2026
Close-up of a hand playing the classic board game Operation, reaching to remove a piece from the patient's body on the yellow game board.
Engineering Guides

How to Drop Noisy Health Check Logs Before They Hit Your Observability Platform

Healthcheck logs generate millions of identical lines per day and silently inflate your observability bill, but filtering them at the right layer can cut total log volume by 15 to 40 percent.
April 7, 2026
Battle image of Grepr and New Relic, with a lightning bolt in between
Comparisons

New Relic Pipeline Control vs Grepr: Manual Rules vs AI Automation

New Relic Pipeline Control bills you on data volume before any filtering happens, requires manual YAML config for every pipeline, and needs a separate Kubernetes install per environment.
April 2, 2026

Get started free and see Grepr in action in 20 minutes.