There’s No “Data Lakehouse” Without Interactive Analytics
There’s No “Data Lakehouse” Without Interactive Analytics
5 Min Read

Before we take a deeper look at data lakehouse solutions, let’s take a step back and start from the very beginning. It wasn’t long ago when data lakes (sometimes also called data hubs) were considered an emerging, bleeding-edge approach to collecting and managing semi-structured and unstructured data. Today, data lakes are in the mainstream, with 58% adoption per a recent 451 Research survey.

Why the rapid rise in popularity? Because although data lakes were originally premised on relatively crude data processing with MapReduce, that vision evolved over time as distributed SQL query engines like Apache Hive, Apache Impala, or Presto appeared to help you get interactive analytics out of your data lake engine via interactive SQL queries. There was even talk (gasp!) of data lakes replacing data warehouses.

Interactive Analytics: Promises Made, but Not Kept

The intricacies of interactive analytics are highly complicated and sophisticated. However, the core purpose of this concept — i.e. the answer to the fundamental question “what is interactive analytics?” — is that it must rapidly scrutinize, analyze, organize, and ultimately transform an enormous volume of raw information into reliable, actionable business intelligence. It must turn noise into substance. Well, it didn’t work out that way.

Today, plentiful research confirmed by Yellowbrick customers tells us that data lake solutions are endlessly frustrating because they only deliver on half their promise: While on-premises HDFS and cloud object stores are great options for low-cost storage, a data lake engine fails to meet enterprise needs for ad hoc, interactive queries for more than a relatively small number of concurrent users. That’s why we’re seeing new patterns like the “data lakehouse” emerge, with the objective of bringing data warehouse-level query latency and concurrency to data lakes.

What is a Data Lakehouse?

In essence, a data lakehouse is a solution that combines the management features and tools of a data warehouse (which are needed to implement governance and schema), with the flexible, low-cost storage of unstructured data of a data lake. In this sense, we could answer the question “what is a data lakehouse?” by saying that it represents a re-imagination and evolution of conventional data warehouses, which have been around for several decades.

Augmenting Data Lakehouse Architecture with Rich Workload Management 

Across the industry, though, we don’t all agree on how to bring those attributes to data lakes. At Yellowbrick, we don’t think customers are well served by caching strategies that try to simulate data warehouse performance but that fall down under the pressure of mixed workloads and high concurrency. Instead, we believe the right answer is to augment your data lakehouse architecture with a modern, real-time SQL data warehouse with rich workload management capabilities, that makes ingestion from HDFS and cloud object stores agile and easy, that has an efficient storage and access format for analyzing structured data, and that’s battle-tested for superb price/performance at scale on real-time and at-rest data. The answer is not to build on solutions that are known to add more cost and complexity for customers; the industry tried that in the past with Hadoop, and we know how that story ends.

Breaking the Limitations

Sounds like a simple goal, but the reality is that current options for reaching it are as problematic as running queries on a data lake engine: Legacy approaches are too inflexible for semi- and unstructured data and expensive to own and scale, while cloud-only options have poor or unpredictable price/performance for large-scale data or concurrency needs (and fail to provide an on-premises option for geo-location or other requirements).

To solve that problem, Yellowbrick has invested years of R&D in an MPP data warehouse designed from the ground up to take advantage of optimized instances for ad hoc, interactive analytics at any scale. Yellowbrick Data Warehouse is a hybrid row- and column-oriented implementation, offering the scan throughput of column stores along with real-time, row-by-row ingestion of data via Apache Kafka (millions of rows/sec). With this approach, streaming data can be ingested and queried alongside at-rest data immediately, with no partitions, manual indexes, query tuning, or other optimizations needed—it just works!

data lakehouse architecture

Yellowbrick’s advanced workload management is also a critical component that prevents resource contention by letting you prioritize queries for specific workload types and/or consumers—for example, you may want to prioritize ad hoc queries from the data science team—while penalty-boxing or even re-starting failed ones. To use a traffic engineering metaphor, for maximum traffic throughput, it’s not enough that the roadway is wide and clear of obstructions; sometimes, you also have to prevent slower vehicles that tie up traffic from entering the road.

Finally, unlike cloud-only options, Yellowbrick data warehouses can run on-premises, in the cloud, and at the network edge.

Customer Results Tell the Story

Customers have already seen the light about the benefits of Yellowbrick’s architecture and workload management capabilities for interactive queries on data lakes. For example, the ThreatMetrix solution from LexisNexis is an industry-leading SaaS product focusing on digital identity verification with more than 5,000 brands as customers. For that use case, poor customer experience has direct business implications. After replacing Impala with Yellowbrick to augment its data lake, ThreatMetrix now ingests 1TB+ of streaming data into the database daily for immediate analysis, user queries that ran in minutes now complete in seconds or less (with ¼ of the infrastructure), and downtime is no longer a concern.

Similarly, one of the biggest mobile operators in Latin America faced a data crisis when its legacy data lake augmentation platform was failing in the face of an 800% growth in data volume. With Yellowbrick, operational reports are now up-to-the-second accurate, revenue reconciliation is done in real time instead of batch (safeguarding millions from SIM card resellers), and queries run 20x faster on 8x more data.

Data Lake Solutions to Consider

Data lakes cannot replace a data warehouse. Successful data lakes need a data warehouse. Data can’t always be centralized into one single data warehouse or data lake. By embracing data dispersed across the enterprise, organizations can bring together and analyze vast quantities data in AWS, Azure, GCP, on-premises and even the edge, in real time while eliminating cloud concentration risk.

As you can see, there’s no data lakehouse without interactive analytics and the workload management that makes it happen. Request a free Test Drive to get hands-on for learning more!

Sign up for our newsletter and stay up to date