

(kurhan/Shutterstock)
The concept of “observability” is well understood as it pertains to DevOps and site reliability engineering (SRE). But what does it mean in the context of data? According to Barr Moses, the CEO of data observability startup Monte Carlo, it’s all about being able to trust the data.
In the last decade or so, we’ve gotten really good at aggregating, storing, collecting, and analyzing data, Moses says. We’re able to move vast amounts of data from sources into data warehouses and data lakes, and utilize the data for dashboards and machine learning models.
“But whether we can actually trust the data is something we haven’t figured out yet,” Moses tells Datanami. “In companies, small to large, really the worst thing that can happen is when you start using the data, but it actually can’t be trusted.”
In fact, making decisions based on bad data can actively hurt a company. Considering all of the ways that humans and algorithms are consuming data in business today, using bad data is worse than doing nothing at all.
There are many ways that data can go bad. From data-entry or coding errors to malfunctioning sensors and data drift, the sources of data contamination are numerous. We can’t guarantee that data will never go bad, so the next best thing is to detect the bad data as soon as possible.
That’s the idea behind data observability. According to Moses, data observability borrows well-established observability concepts from DevOps and SRE in one key respect: the importance of monitoring the outputs of data systems to determine if something is going awry inside the system.
“If you think what DevOps or software engineer teams are tasked with, they have many applications and system infrastructure that they are tasked with making sure that they are up and running at all times,” Moses says. “It’s a very well-understood approach in DevOps, but it’s completely new to data. That’s what we call data observability.”
Monte Carlo’s data observability offering is structured around five pillars of observability:
- Freshness, or the timeliness of the data;
- Volume, of the completeness of the data;
- Distribution, which measures the consistency of data at the field level;
- Schema, relating to the structure of fields and tables;
- Lineage, or a change-log of the data.
“The only way to truly solve data observability is to do that in an end-to-end way,” Moses says. “So it includes the customers’ entire data stack. That includes cloud data lakes, data warehouse, ETL, BI, and machine learning models.”
The company, which has raised $41 million in venture capital to date, built connectors that pull data from each component of the stack, and monitors the data in a read-only manner. If the output begins to show signs of a problem – such as a dashboard reporting null values – the software will automatically generate an alert and send it via email, text, Slack or PagerDuty.
It’s about avoiding data downtime. Just as DevOps and SRE teams have instrumented their systems in an attempt to detect the faintest hint of a pending failure that could take production systems offline, data observability, as Monte Carlo practices it, takes a holistic approach to monitoring a range of data characteristics to figure out when good data is breaking bad.
Moses draws a parallel between what Monte Carlo is setting out to do and what New Relic, DataDog, and App Dynamics are doing in the DevOps and SRE space. While data infrastructure providers (i.e. vendors building databases, data lakes, data warehouses and ETL, BI, and ML tools) supply some monitoring capabilities for their products, it’s an industry best practice to tap third-party observability and monitoring tools to bring all of those components together into a comprehensive view.
“I think it’s important to have a new layer in the modern data stack,” says Moses, who was a management consultant for Bain & Company before founding Monte Carlo with Lior Gavish in 2019. “And part of the new layer is data observability. It’s important for it to be a third-party that can be integrated with all of these vendors and solutions…and also provide a third-party objective view” that can be trusted.
Today, the San Francisco company announced a partnership with cloud data warehouse provider Snowflake. The deal will see Monte Carlo become a “native” provider of data observability for Snowflake customers.
“Data is only powerful if you can actually trust it,” Moses says. “We firmly belief that’s a key part to building a strong data platform and data architecture. The Snowflake team has been really supportive of our vision and this new category of data observability. We’re exited to partner with them to bring this to the our customers.”
Related Items:
Do You Have Customer Data You Can Trust?
July 30, 2025
- Elastic Announces Faster Filtered Vector Search with ACORN-1 and Default Better Binary Quantization Compression
- Nutanix Named a Leader in Multicloud Container Platforms Evaluation
- RAVEL Expands Orchestrate AI With Supermicro-Based AI Workload Solution
- MLCommons Releases MLPerf Client v1.0: A New Standard for AI PC and Client LLM Benchmarking
- IBM: 13% of Organizations Reported Breaches of AI Models, 97% of Which Reported Lacking Proper AI Access Controls
- Hitachi Vantara Announces Virtual Storage Platform One for Hybrid Cloud Data Management
- Elastic Delivers New ES|QL Features for Cross-Cluster Scale, Data Enrichment, and Performance
- Cognizant Launches AI Training Data Services to Accelerate AI Model Development at Enterprise Scale
- Fractal Launches Agentic AI Platform Cogentiq to Drive Enterprise Performance
July 29, 2025
- Git-for-data Pioneer lakeFS Secures $20M in Growth Capital, Fills a Critical Gap in Enterprise AI Tech Stack
- Esri, Microsoft, and Space42 Partner to Launch ‘Map Africa Initiative’
- Teradata Expands ModelOps in ClearScape Analytics for Generative and Agentic AI
- Linux Foundation Welcomes AGNTCY to Tackle AI Agent Fragmentation
- Deloitte: Trust Emerges as Main Barrier to Agentic AI Adoption in Finance and Accounting
- Lightbits Launches NVMe over TCP Storage for Kubernetes on Supermicro Systems, Unveiling Benchmark Results
- AWS and dbt Labs Sign Strategic Collaboration Agreement
- Actian Study Finds Organizations Overestimate Data Governance Maturity, Posing Risk to AI Investments
- Privacera Named Leader in GigaOm Radar for Data Access Governance for 4th Consecutive Time
July 28, 2025
- Scaling the Knowledge Graph Behind Wikipedia
- LinkedIn Introduces Northguard, Its Replacement for Kafka
- Top 10 Big Data Technologies to Watch in the Second Half of 2025
- Iceberg Ahead! The Backbone of Modern Data Lakes
- What Are Reasoning Models and Why You Should Care
- Apache Sedona: Putting the ‘Where’ In Big Data
- Top-Down or Bottom-Up Data Model Design: Which is Best?
- Rethinking AI-Ready Data with Semantic Layers
- What Is MosaicML, and Why Is Databricks Buying It For $1.3B?
- Rethinking Risk: The Role of Selective Retrieval in Data Lake Strategies
- More Features…
- Supabase’s $200M Raise Signals Big Ambitions
- Mathematica Helps Crack Zodiac Killer’s Code
- Solidigm Celebrates World’s Largest SSD with ‘122 Day’
- AI Is Making Us Dumber, MIT Researchers Find
- Promethium Wants to Make Self Service Data Work at AI Scale
- The Top Five Data Labeling Firms According to Everest Group
- Toloka Expands Data Labeling Service
- With $20M in Seed Funding, Datafy Advances Autonomous Cloud Storage Optimization
- Ryft Raises $8M to Help Companies Manage Their Own Data Without Relying on Vendors
- AWS Launches S3 Vectors
- More News In Brief…
- Seagate Unveils IronWolf Pro 24TB Hard Drive for SMBs and Enterprises
- Gartner Predicts 40% of Generative AI Solutions Will Be Multimodal By 2027
- OpenText Launches Cloud Editions 25.3 with AI, Cloud, and Cybersecurity Enhancements
- TigerGraph Secures Strategic Investment to Advance Enterprise AI and Graph Analytics
- Promethium Introduces 1st Agentic Platform Purpose-Built to Deliver Self-Service Data at AI Scale
- StarTree Adds Real-Time Iceberg Support for AI and Customer Apps
- Databricks Announces Data Intelligence Platform for Communications
- Gathr.ai Unveils Data Warehouse Intelligence
- Graphwise Launches GraphDB 11 to Bridge LLMs and Enterprise Knowledge Graphs
- Campfire Raises $35 Million Series A Led by Accel to Build the Next-Generation AI-Driven ERP
- More This Just In…