Overview
Introduction
SulacoTec operates in the click-brokering layer of the ad stack: ingesting bid
requests, buying traffic in bulk, and reselling performance to downstream customers.
Historically, bid prices were tuned using heuristics and small samples of performance
data, which made it hard to price large-volume deals with confidence.
As a Senior Software Engineer, I was tasked with turning a large historical
click-processing database into a decision engine. The goal was to predict reasonable
CPC, CPA, and CPM ranges for given traffic segments so the business could negotiate
and execute bulk click purchases with more predictable margins.
Background
Context
Over time, SulacoTec accumulated ~100 million rows of historical data across:
- Bid requests and clearing prices.
- Clicks, conversions, and post-click outcomes.
- Traffic sources, geos, devices, and creative metadata.
However, this data mostly sat in operational tables used for reporting, not as an
input to predictive decisions. There was a clear opportunity: use the historical
record to estimate expected performance for new campaigns and traffic types.
Challenge
Problem
The business needed to answer a seemingly simple question:
“If we buy N clicks from this traffic source and geo, what CPC/CPA/CPM should we
target to stay profitable for a given customer?”
In practice, this meant:
- Dealing with noisy, sparse, and sometimes inconsistent historical data.
- Capturing non-obvious relationships (e.g., device × geo × time-of-day).
- Delivering answers quickly enough to be useful in negotiations and campaign setup.