RESEARCH & RESOURCES

TDWI Checklist Report | Preparing Your Organization for In-Memory Computing

August 7, 2014

The “decision support system” concept that emerged more than 20 years ago has evolved into the modern data warehouse (DW) that is now ubiquitous in any enterprise information architecture. However, early designers, driven by the need to preserve online transaction processing (OLTP) response times, made a practical engineering decision to replicate transaction data to a segregated platform that would house the reporting and business intelligence (BI) systems.

The conventional wisdom for BI and analytics insists on an intricate infrastructure that iteratively extracts source data sets, applies transformations, and loads the data into a segregated environment. The benefit of system segregation to protect performance is offset by the overhead and complexity associated with data integration, which in turn has contributed to skyrocketing costs for acquiring and maintaining the data warehouse ecosystem’s hardware and software.

Fortunately, in-memory database management systems promise accelerated application performance balanced by increased systemic simplicity. Combining alternative storage layouts with in-memory processing allows applications to take advantage of efficient use of available memory and reduce or even eliminate the data latencies typically associated with significantly slower disk-based storage media.

If reporting and analytics systems are still dissociated from data sources, the performance enhancements gained by keeping the data in main memory diminish as extraction and loading latencies are reintroduced. However, as RAM and cache memory costs plummet and massive multiprocessor systems enter the mainstream, it makes sense to leverage in-memory computing systems such as SAP HANA to fully integrate transaction processing, operational processing, and reporting and analytics within the same in-memory platform.

This TDWI Checklist Report will help your organization prepare for in-memory computing by discussing how it improves application performance. In-memory computing enables the peaceful coexistence of transaction processing and analytics without demanding data extraction, transformations, rampant replication, and heterogeneous computing platforms. We examine the impact of in-memory computing on the corporate data footprint and how the information technology landscape is made simpler.

A transition to in-memory computing makes information management easier by freeing hardware and staff resources for productive use, inspiring disruptive innovation that enables easier application development, improved user experience, and general increase in corporate value.

Your e-mail address is used to communicate with you about your registration, related products and services, and offers from select vendors. Refer to our Privacy Policy for additional information.


TDWI Membership

Get immediate access to training discounts, video library, BI Teams, Skills, Budget Report, and more

Individual, Student, & Team memberships available.