Data Pipeline Management and Security Data Fabrics

Data Pipeline Management and Security Data Fabrics In the recent past, DataBahn has been featured in 3 different narratives focused on security data – Being mentioned by these experts is a welcome validation. It is also a recognition that we are solving a relevant problem for businesses – and for these mentions to come from...

November 14, 2024
|

Data Pipeline Management and Security Data Fabrics

In the recent past, DataBahn has been featured in 3 different narratives focused on security data -

  1. Cybersecurity experts like Cole Grolmus (Strategy of Security) discussing how DataBahn's "Security Data Fabric" solution is unbundling security data collection from SIEMs (Post 1, Post 2)
  2. VCs such as Eric Fett of NGP Capital talking about how DataBahn's AI-native approach to cybersecurity was revolutionizing enterprise SOCs attempts to combat alert fatigue and escalating SIEM costs (Blog)
  3. Most recently, Allie Mellen, a Principal Analyst at Forrester, shouted DataBahn out as a "Data Pipeline Management" product focusing on security use cases. (LinkedIn Post, Blog)

Being mentioned by these experts is a welcome validation. It is also a recognition that we are solving a relevant problem for businesses – and for these mentions to come from these different sources represents the perspectives from which we can consider our mission.

What are these experts saying?

Allie’s wonderful article (“If You’re Not Using Data Pipeline Management for Security and IT, You Need To”) expertly articulates how SIEM spending is increasing, and that SIEM vendors haven’t created effective tools for log size reduction or routing since it “… directly opposes their own interests: getting you to ingest more data into their platform and, therefore, spend more money with them.”

This aligns with what Cole alluded to, when he stated reasons why "Security Data Fabrics” shouldn’t be SIEMs", pointing to this same conflict of interest. He added that these misaligned incentives spilled over into interoperability, where proprietary data formats and preferred destinations would promote vendor lock-in, which he had previously mentioned Security Data Fabrics were designed to overcome.

Eric’s blog was focused on identifying AI-native cybersecurity disrupters, where he identified DataBahn as one of the leading companies whose architecture was designed to leverage and support AI features, enabling seamless integration into their own AI assets to “ … overcome alert fatigue, optimize ingestion costs, and allocate resources to the most critical security risks.”

What is our point of view?

The reflections of these experts resonate with our conception of the problem we are trying to solve—SOCs and Data Engineering teams overwhelmed by the laborious task of data management and the prohibitive cost of the time and effort involved in overcoming it.

  • SIEM ingest costs are too high. ~70% of the data being sent to SIEMs is not security-relevant. Logs have extra fields you don’t always need, and indexed data becomes 3-5x the original size. SIEM pricing data depends upon the volume of data being ingested and stored with them – which strains budgets and reduces the value that SOCs get from their SIEMs.

    We deliver a 35%+ reduction in SIEM costs by reducing log sizes in 2-4 weeks – and our AI-enabled platform enables ongoing optimization to continue to reduce log sizes.
  • SIEM being the source of data ingestion is also a problem. SIEMs are not very good at data ingestion. While some SIEM vendors have associated cloud environments (Sentinel, SecOps) with native ingestion tools, adding new sources – especially custom apps or sources with unstructured data – requires extensive data engineering effort and 4-8 weeks of integration. Additionally, managing these data pipelines is challenging, as these pipelines are single points of failure. Back pressure and spikes in data volumes can cause data loss.

    DataBahn ensures loss-less data ingestion via a mesh architecture that ensures a secondary channel to ingest data in case of any blockage. It also tracks and identifies sudden changes in volume, helping to identify issues faster.
  • Data Formats and Schemas are a challenge. SIEMs, UEBAs, Observability Tools, and different data storage destinations come with their proprietary formats and schemas, which add another manual task of data transformation onto data engineering teams. Proprietary formats and compliance requirements also create vendor lock-in situations, which add to your data team’s cost and effort.

    We automate data transformation, ensuring seamless and effortless data standardization, data enrichment, and data normalization before forking the data to the destination of your choice.

Our solution is designed for specific security use cases, including a library of 400+ connectors and integrations and 900+ volume reduction rules to reduce SIEM log volumes, as well as support for all the major formats and schemas – which puts it ahead of generic DPM tools, something which Allie describes in her piece.

Cole has been at the forefront of conversations around Security Data Fabrics, and has called out that DataBahn has built the most complete platform/product in the space, with capabilities across integration & connectivity, data handling, observability & governance, reliability & performance, and AI/ML support.

Conclusion

We are grateful to be mentioned in these vital conversations about security data management and its future, and we appreciate the time and effort being spent by these experts to drive these conversations. We hope that this increases the awareness of Data Pipeline Management, Security Data Fabrics, and AI-native data management tools - a venn diagram we are pleased to be at the intersection of - and look forward to continue our journey in solving the problems that these experts have identified.

Uncover hidden visitor insights to improve their website journey
Share

See related articles

Why is DataBahn building agents? Why now?  

Agents are not new. But the problem they were created to solve has evolved. What’s changed is not just the technology landscape, but the role of telemetry in powering modern detection, response, AI analytics, and compliance. 

Most endpoint agents were designed for a narrow task: collect logs, ship them somewhere, and stay out of the way. But today’s security pipelines demand more. They need selective, low-latency, structured data that feeds not just a SIEM, but an entire ecosystem, from detection engines and data lakes to streaming analytics and AI models. 

Our mission has always been to eliminate data waste and simplify how enterprises move, manage, and monitor security data. That’s why we built the Smart Agent: a lightweight, programmable collection layer that brings policy, precision, and platform awareness to endpoint telemetry – without the sprawl, bloat, and hidden costs of traditional agents. 

A Revolutionary Approach to Endpoint Telemetry

Traditional agents are often built as isolated tools – one for log forwarding, another for EDR, a third for metrics. This results in resource contention, redundant data, and operational sprawl. 

DataBahn's Smart Agent takes a fundamentally different approach. It’s built as a platform-native component, not a point solution. That means collect once from the endpoint, normalize once, and route anywhere, breaking the cycle of duplication.  

Here’s what sets it apart: 

- Modular, Policy-Driven Control: Enterprise teams can now define exactly what to collect, how to filter or enrich it, and where to send it – with full version control, change monitoring, and audit trails.  

- Performance Without Sprawl: Replace 3–5 overlapping agents per endpoint with a single lightweight Smart Edge agent that serves security, observability, and compliance workflows simultaneously.  

- Built for High-Value Telemetry: Our agents are optimized to selectively capture only high-signal events, reducing compute strain and downstream ingestion costs.  

- AI-Ready, Future-Proof Architecture: These agents are telemetry-aware and natively integrated into our AI pipeline. Whether it’s streaming inference, schema awareness, or tagging sensitive data for compliance – they’re ready for the next generation of intelligent data pipelines.  

This isn’t just about replacing old agents. It’s about rethinking the endpoint as the first intelligent node in your data pipeline. 

Solving Real Enterprise Problems 

We’ve spent years embedded in complex environments – from highly regulated banks to fast-moving cloud-native tech firms. And across the board, one pattern kept surfacing: traditional approaches to endpoint telemetry don’t scale. 

  • Agent Sprawl is Draining Resources: Too many agents, too much overhead. Each one comes with its own update cycles, configuration headaches, and attack surface. Our agents consolidate that complexity – offering centralized control, real-time health monitoring, and zero-downtime updates. 
  • Agentless Left Security Teams in the Dark: APIs and control planes can’t capture runtime behavior, memory state, or user actions in real time. Our agents plug that gap – giving enterprises low-latency, high-fidelity data from endpoints, VMs, containers, and edge devices. 
  • Latency, Duplication, and Blind Spots: Polling intervals and subscription models delay detection. Meanwhile, multiple agents flood SIEMs with duplicate telemetry. DataBahn's agents are event-driven, deduplicated, and volume-aware – reducing noise and improving signal quality. 
  • A Platform Approach to Edge Data: DataBahn’s agents are not just better versions of old tools – they represent a strategic shift: a unified data layer from endpoint to cloud, where telemetry is no longer hardcoded to tools, vendors, or formats. 

What that enables: 

  • Multiple Deployment Models: Direct-to-destination, hybrid agentless, or agent-per-asset based on asset value. 
  • Seamless integration with our Smart Edge: Making it easy to extend telemetry pipelines, apply real-time transformations, and deliver enriched data to multiple destinations – without code. 
  • Compliance-Ready Logging: Built-in support for log integrity, masking, and tagging to meet industry standards like PCI, HIPAA, and GDPR. 

The End of the Agent vs. Agentless Debate  

The conversation around data collection has been stuck in a binary: agent or agentless. But in real-world environments, that framing doesn’t hold.  

What enterprises need isn’t one or the other but the ability to deploy the right mechanism based on asset type, risk, latency sensitivity, and the downstream use case.  

The future isn’t agent or agentless – it’s context-aware, modular, and unified. Data collection that adapts to where it’s running, integrates cleanly into existing pipelines, and remains extensible for what comes next, whether that’s AI-driven security operations, privacy-focused compliance, or cross-cloud observability.  

That’s the shift we’re enabling with the DataBahn Smart Agent. Not just a product – but a programmable foundation for secure, scalable, and future-ready telemetry.

In their article about how banks can extract value from a new generation of AI technology, notable strategy and management consulting firm McKinsey identified AI-enabled data pipelines as an essential part of the ‘Core Technology and Data Layer’. They found this infrastructure to be necessary for AI transformation, as an important intermediary step in the evolution banks and financial institutions will have to make for them to see tangible results from their investments in AI.

The technology stack for the AI-powered banking of the future relies greatly on an increased focus on managing enterprise data better. McKinsey’s Financial Services Practice forecasts that with these tools, banks will have the capacity to harness AI and “… become more intelligent, efficient, and better able to achieve stronger financial performance.

What McKinsey says

The promise of AI in banking

The authors point to increased adoption of AI across industries and organizations, but the depth of the adoption remains low and experimental. They express their vision of an AI-first bank, which -

  1. Reimagines the customer experience through personalization and streamlined, frictionless use across devices, for bank-owned platforms and partner ecosystems
  2. Leverages AI for decision-making, by building the architecture to generate real-time insights and translating them into output which addresses precise customer needs. (They could be talking about Reef)
  3. Modernizes core technology with automation and streamlined architecture to enable continuous, secure data exchange (and now, Cruz)

They recommend that banks and financial service enterprises set a bold vision for AI-powered transformation, and root the transformation in business value.

AI stack powered by multiagent systems

The true potential of AI will require banks of the future to tread beyond just AI models, the authors claim. With embedding AI into four capability layers as the goal, they identify ‘data and core tech’ as one of those four critical components. They have augmented an earlier AI capability stack, specifically adding data preprocessing, vector databases, and data post-processing to create an ‘enterprise data’ part of the ‘core technology and data layer’. They indicate that this layer would build a data-driven foundation for multiple AI agents to deliver customer engagement and enable AI-powered decision-making across various facets of a bank’s functioning.

Our perspective

Data quality is the single greatest predictor of LLM effectiveness today, and our current generation of AI tools are fundamentally wired to convert large volumes of data into patterns, insights, and intelligence. We believe the true value of enterprise AI lies in depth, where Agentic AI modules can speak and interact with each other while automating repetitive tasks and completing specific and niche workstreams and workflows. This is only possible when the AI modules have access to purposeful, meaningful, and contextual data to rely on.

We are already working with multiple banks and financial services institutions to enable data processing (pre and post), and our Cruz and Reef products are deployed in many financial institutions to become the backbone of their transformation into AI-first organizations.

Are you curious to see how you can come closer to building the data infrastructure of the future? Set up a call with our experts to see what’s possible when data is managed with intelligence.

Two years ago, our DataBahn journey began with a simple yet urgent realization: security data management is fundamentally flawed. Enterprises are overwhelmed by security and telemetry, struggling to collect, store, and process it, while finding it harder and harder to gain timely insights from it. As leaders and practitioners in cybersecurity, data engineering, and data infrastructure, we saw this pattern everywhere: spiraling SIEM costs, tool sprawl, noisy data, tech debt, brittle pipelines, and AI initiatives blocked by legacy systems and architectures.

We founded DataBahn to break this cycle. Our platform is specifically designed to help enterprises regain control: disconnecting data pipelines from outdated tools, applying AI to automate data engineering, and constructing systems that empower security, data, and IT teams. We believe data infrastructure should be dynamic, resilient, and scalable, and we are creating systems that leverage these core principles to enhance efficiency, insight, and reliability.

Today, we’re announcing a significant milestone in this journey: a $17M Series A funding round led by Forgepoint Capital, with participation from S3 Ventures and returning investor GTM Capital. Since coming out of stealth, our trajectory has been remarkable – we’ve secured a Fortune 10 customer and have already helped several Fortune 500 and Global 200 companies cut over 50% of their telemetry processing costs and automate most of their data engineering workloads. We're excited by this opportunity to partner with these incredible customers and investors to reimagine how telemetry data is managed.

Tackling an industry problem

As operators, consultants, and builders, we worked with and interacted with CISOs across continents who complained about how they had gone from managing gigabytes of data every month to being drowned by terabytes of data daily, while using the same pipelines as before. Layers and levels of complexity were added by proprietary formats, growing disparity in sources and devices, and an evolving threat landscape. With the advent of Generative AI, CISOs and CIOs found themselves facing an incredible opportunity wrapped in an existential threat, and without the right tools to prepare for it.

DataBahn is setting a new benchmark for how modern enterprises and their CISO/CIOs can manage and operationalize their telemetry across security, observability, and IOT/OT systems and AI ecosystems. Built on a revolutionary AI-driven architecture, DataBahn parses, enriches, and suppresses noise at scale, all while minimizing egress costs. This is the approach our current customers are excited about, because it addresses key pain points they have been unable to solve with other solutions.

Our two new Agentic AI products are solving problems for enterprise data engineering and analytics teams. Cruz automates complex data engineering tasks from log discovery, pipeline creation, tracking and maintaining telemetry health, to providing insights on data quality. Reef surfaces context-aware and enriched insights from streaming telemetry data, turning hours of complex querying across silos into seconds of natural-language queries.

The Right People

We’re incredibly grateful to our early customers; their trust, feedback, and high expectations have shaped who we are. Their belief drives us every day to deliver meaningful outcomes. We’re not just solving problems with them, we’re building long-term partnerships to help enterprise security and IT teams take control of their data, and design systems that are flexible, resilient, and built to last. There’s more to do, and we’re excited to keep building together.

We’re also deeply thankful for the guidance and belief of our advisors, and now our investors. Their support has not only helped us get here but also sharpened our understanding of the opportunity ahead. Ernie, Aaron, and Saqib’s support has made this moment more meaningful than the funding; it’s the shared conviction that the way enterprises manage and use data must fundamentally change. Their backing gives us the momentum tomove faster, and the guidance to keep building towards that mission.

Above all, we want to thank our team. Your passion, resilience, and belief in what we’re building together are what got us here. Every challenge you’ve tackled, every idea you’ve contributed, every late night and early morning has laid the foundation for what we have done so far and for what comes next. We’re excited about this next chapter and are grateful to have been on this journey with all of you.

The Next Chapter

The complexity of enterprise data management is growing exponentially. But we believe that with the right foundation, enterprises can turn that complexity into clarity, efficiency, and competitive advantage.

If you’re facing challenges with your security or observability data, and you’re ready to make your data work smarter for AI, we’d love to show you what DataBahn can do. Request a demo and see how we can help.

Onwards and upwards!

Nanda and Nithya
Cofounders, DataBahn