The Infrastructure Problem: Customized Pipelines and Guide Workflow
Superhuman runs one of many largest AI productiveness platforms on the earth, processing textual content throughout dozens of languages for tens of millions of customers on daily basis. A unified knowledge platform on Databricks powers analytics, machine studying and the pipelines behind customer-facing options. Nevertheless, the infrastructure connecting that knowledge to manufacturing functions and go-to-market workflows had grown fragile, creating a major engineering burden on two fronts.
How a Redis sync failure uncovered Superhuman’s pipeline fragility
On the infrastructure facet, the ML staff had constructed a customized pipeline to sync knowledge from the Databricks Lakehouse into Redis and DynamoDB. The pipeline powered eligibility guidelines for promotions and free trials, issues like whether or not a person had activated a particular characteristic or logged sufficient periods to qualify for a reduction. “It was the precise answer when it was constructed. However priorities shifted, the staff moved on to different issues, and the system stayed behind,” stated Michael Kobelev, ML Infra Software program Engineer. Not too long ago, a routine library replace broke sync jobs writing to Redis. With out centralized alerting, the staff could not establish all of the affected jobs and a few failures did not floor till weeks later.
The last-mile hole between warehouse knowledge and gross sales workflows
On the go-to-market facet, the hole between knowledge and motion was simply as expensive. Gross sales and buyer success reps manually copied metrics from dashboards into PowerPoint templates, spending roughly half-hour per deck, a number of instances every week. “Knowledge groups construct unimaginable pipelines and tables, however there’s nonetheless a final mile between what’s within the warehouse and what a gross sales rep can really use,” stated Maximilian Proano, Software program Engineer, Knowledge Purposes. When the staff constructed Deckster, an LLM-powered app that mechanically generates customer-facing shows, the primary model pulled metrics dwell from a SQL warehouse. Latency spiked on each click on, and there was no clear method to cache outcomes or persist a person’s progress.
Each groups wanted the identical factor. A managed transactional layer that sat near their Delta tables and required minimal repairs.





