Why I joined Molecula…
The world of data has been evolving for decades, and organizations continually struggle to actually get the impact they hope to get from the data they collect. This problem is exacerbated by large, disparate data sets or the need for real-time insights on high-velocity, inbound data. In fact, according to McKinsey Insights, organizations are currently making decisions based on analyzing just 1% of the data in their possession. The opportunity to use ALL the data available and not be constrained by speed, size, access, security, computational complexity, etc. is profound for virtually every industry. Some are calling it the ‘super evolution.’
As we enter into the next evolutionary phase, humans and machines will need access to vast amounts of data as fluidly as the speed of thought in order to survive and ultimately thrive. Today’s technology has done a great job of data collection (i.e. with Hadoop and data lakes, data pipelines like Kafka, clouds, etc.). However, we still have not figured out the best approach to data access at the scale, speed, security, and compliance required for true machine learning, artificial intelligence or IoT use cases. These use cases require massive amounts of data, often in real time, residing in various different locations and stored in different formats, crossing different clouds, countries/jurisdictions, and edge locations.
I’ve Seen the Problems
Many companies are using incremental fixes to attempt a solution with tactics such as query federation, aggregation and caching (making copies of values), copy management, or hardware-based acceleration. However, none of these options are truly groundbreaking enough to fundamentally enable organizations to get us to the orders of magnitude faster, more secure data access that is required.
I have been part of the analytics, BI, Big Data and IoT space for 25+ years. In that time, I always fundamentally believed that there had to be a better way; it just hadn’t been invented yet. I will never forget being at SAP Sapphire in 2010 when Hasso Platner and Vishal Sikka stood on stage and introduced SAP HANA—the first in-memory database that would enable decision-making insights for organizations that would be orders of magnitude faster and cheaper, across every industry. I got chills thinking about how this would impact our ability to use data to solve any problem. However, just a decade later as we begin to move past the Information Era, we have surpassed even SAP HANA to answer our next-generation data usage needs. So what is the next disruption that will usher us into the Intelligence Era?
How I See Molecula
I fundamentally believe Molecula is the next generation data access layer for systems of engagement. When I learned what Molecula was doing, I had the same feeling I had ten years ago at SAP Sapphire. I realized this could be the building block to truly unlock ALL data at the speed, cost, and governance necessary in the AI, ML, and IoT era. Molecula will enable teams of humans and machines to make decisions orders of magnitude faster and cheaper.
Imagine the ability to:
- know your customers’ behavior and personalize their experience in real time
- optimize machine learning models in real time, leveraging historic and current machine data with no data or time limitations
- make instant decisions on the edge based on all edge devices across tens, thousands, or even tens of thousands of locations
Molecula is an enterprise-grade, zero-copy Data Virtualization platform that enables insights to flow at the speed of thought across all of your data with no compromise of security, compliance, speed, size, data type, location, or format. Molecula does this by creating a mathematical abstraction layer of any given data source, irrespective of size or location, that represents the underlying data but is not actually a copy. Without moving or caching the data, it stores this unified access layer in a format that is extremely compact and portable called the Virtual Data Source (VDS). The VDS can span across silos of data, organizations, or ecosystems and enable an enterprise-wide unified access layer that becomes a very nimble, secure, compliant, and fast way to access a complete data set.
Molecula’s zero-copy, zero-moving, zero-caching data virtualization approach gets you from data to decision without the typical aggregations, federations, caching or other techniques employed by information era vendors. While there are some well-known approaches to Data Virtualization in the market today, we believe none will suffice in the next generation of data usage, and none compare to the Molecula technology.
The Power of True Data Virtualization
“Data Virtualization was invented to provide enterprises a centralized and singular view of their disparate data sources. Through 2022, 60% of all organizations will implement Data Virtualization as one key delivery style in their data integration architecture,” according to Gartner. Gartner Market Guide for Data Virtualization, 16 November 2018.
I believe Molecula is a critical technology to access data for the Intelligence Era. I am so blown away by the Molecula technology, I decided to join the company and become the Chief Strategy and Marketing Officer. I am thrilled to be part of the Molecula team and help customers across industries to achieve their full potential with true massive scale and instantaneous data access. I believe Molecula will help save lives, avoid catastrophes, speed innovation, predict the future, shorten response times, reduce risks, reverse productivity erosion, empower researchers, and much more!
This is the first of my Molecula blog series. Stay tuned for future blogs, and learn how Molecula was invented, how customers are using Molecula and further explore why other methods of creating a ‘unified access layer’ will not suffice in the Intelligence Era.