At our 8th annual
At the analytics breakout session I had the opportunity to sit around a table with CIOs from a variety of insurers of different sizes and lines. One thing that struck me was how much big data and big data technology is part of the conversation. Novarica typically urges caution when asked about big data, stressing that insurers shouldn’t go looking for a use case to fit a technology. But it seems that more insurers are finding the use case which leads them towards a big data approach, and it raises questions about the future of the enterprise data warehouse.
Building a data warehouse with a single data model that supports the whole business is notoriously difficult, with a success rate that drops as the amount of data and number of lines of business get larger, until becoming a near impossibility. A big data approach allows bringing together many different data sources with different structures without having to go through the normalizing and cleansing process that often derails EDW projects. While both approaches have value, larger insurers are discovering the big data route may be their only option for cross-business analysis. And some smaller insurers see this as a way to rapidly gather and review data as well, often running side-by-side with an existing data warehouse. In fact, in at least one case, an insurer’s big data lake has become the first step in the data workflow, with a data extraction from it feeding the older data warehouse as a way to maintain legacy processes.
We’re currently reaching out to the
This blog entry has been republished with permission.
Readers are encouraged to respond using the “Add Your Comments” box below.
The opinions posted in this blog do not necessarily reflect those of Insurance Networking News or SourceMedia.