Teradata QueryGrid: Beyond Enterprise Data Warehouse - InformationWeek

InformationWeek is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them.Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

Data Management // Software Platforms
03:00 PM
Connect Directly

Teradata QueryGrid: Beyond Enterprise Data Warehouse

Teradata adds unified fabric access to myriad databases and Hadoop, triggering multiple analysis engines with a single query. Say goodbye to yesterday's enterprise data warehouse ideas.

Teradata, the enterprise data warehouse (EDW) company, announced a QueryGrid data-access layer on Monday that can orchestrate multiple modes of analysis across multiple databases plus Hadoop. It's a next step toward Gartner's vision of a logical data warehouse and an acknowledgement that the notion of the EDW has fundamentally changed.

Teradata has already acknowledged the world beyond the EDW with its Unified Data Architecture, which incorporates the Teradata Aster database for data discovery and Hadoop for varied and voluminous data not well suited to relational database management systems (DBMS). The QueryGrid adds a single execution layer that orchestrates analyses across Teradata, Teradata's Asterdata DBMS, Oracle, Hadoop, and, in the future, other databases and platforms. The analysis options include SQL queries, as well as graph, MapReduce, R-based analytics, and other applications.

"Users don't care if information is sitting inside of a data warehouse or Hadoop, and enterprises don't want a lot of data movement or data duplication," said Chris Twogood, Teradata's vice president of product and services marketing. "The QueryGrid gives them a transparent way to optimize the power of different technologies within a logical data warehouse."

[Want more on the emergence of new analysis techniques? Read Merck Optimizes Manufacturing With Big Data Analytics.]

Offering two-way, Infiniband connectivity among data sources, the QueryGrid can execute sophisticated, multi-part analyses. After finding a segment of high-value customers in Teradata, for example, you could push that subset into Hadoop to explore their sentiments as revealed in Twitter and Facebook social comments. Spotting customers likely to churn -- based on negative sentiments -- you could bring that subset into Asterdata, where graph analysis could be used to spot the most influential customers. Voila: you have a list of high-value, well-connected customers that should be included in an anti-churn campaign.

"There are so many specialized engines, so we want to be able to leverage and integrate those while enabling users of the data warehouse to be able to invoke those techniques," Twogood said.

Teradata isn't the only vendor building what Gartner calls the logical data warehouse. Just last month, SAP introduced its Hana In-Memory Data Fabric for federated data access across sources. And since 2009, IBM has offered its DB2 Information Integrator for federated access to multiple data sources. But where these tools are SQL centric, Teradata's differentiator is broader access to a variety of analysis engines.

With an eye toward heterogeneity, Teradata also introduced Teradata 15 on Monday. This DBMS update adds support for JSON data and goes beyond SQL to invoke applications written in Python, Perl, Ruby, R, and other languages to come.

"This gives you the architectural flexibility to separate the presentation layer and the data-analytics layer," said Alan Greenspan, a product marketing manager at Teradata. "Instead of forcing developers to turn to the data warehouse group to do everything in SQL, they can write their own code and execute it in parallel within the database." The approach avoids data movement, data processing on application servers, and other workarounds between web developers and data-management teams. Greenspan said.

Teradata also announced on Monday an upgrade of its flagship hardware platform. The upgrade offers eight times more memory and three times more solid-state drives per rack than what is delivered in the 6700 series introduced 18 months ago. With 512 gigabytes of memory now available per compute node, Teredata's Intelligent Memory feature can now hold more high-demand data in-memory for lightning-fast, RAM-access-speed analysis.

Introduced in 2013, Teradata Intelligent Memory automatically moves high-demand data to the fastest storage choices available while moving low-demand data to the lowest-cost storage options like high-capacity disk drives. It was a response to SAP's Hana in-memory platform and a preemptive move ahead of the Microsoft and Oracle in-memory options being introduced this year.

The biggest news here is clearly the QueryGrid. After years of preaching that everything should go in the enterprise data warehouse, Teradata is acknowledging and embracing a world in which the EDW doesn't have to be the center of analysis.

What do Uber, Bank of America, and Walgreens have to do with your mobile app strategy? Find out in the new Maximizing Mobility issue of InformationWeek Tech Digest.

Doug Henschen is Executive Editor of InformationWeek, where he covers the intersection of enterprise applications with information management, business intelligence, big data and analytics. He previously served as editor in chief of Intelligent Enterprise, editor in chief of ... View Full Bio

We welcome your comments on this topic on our social media channels, or [contact us directly] with questions about the site.
Comment  | 
Print  | 
More Insights
Newest First  |  Oldest First  |  Threaded View
D. Henschen
D. Henschen,
User Rank: Author
4/7/2014 | 5:50:10 PM
Propagating a myth about Hadoop
RDBMS vendors like to say that Hadoop is all about "unstructured" or "semi-structured" information. And their favorite example is Twitter and Facebook comments. That's not the full truth and it's a bit of a left-handed compliment meant to diminish the value of Hadoop. The fact is, Hadoop can handle very structured information that happens to be varied, voluminuous, inconsistent (a.k.a. sparse) or all of the above.

Check out my recent profile of Merck & Co., which combined 16 different types of very structured data to figure out why some batches of a vaccine had high yield rates and other batches had low yield rates. The advantage was being able to "dump everything in a lake" without time-consuming data modeling and ETL work. Within three months Merck was able to cluster and visualize batch yield rates and spot "smoking guns" within 10 years worth of product and manufacturing plant data. The data wasn't actually all that big -- only 1.5 terabytes -- but the ability to bring together a variety of data quickly made all the difference.
InformationWeek Is Getting an Upgrade!

Find out more about our plans to improve the look, functionality, and performance of the InformationWeek site in the coming months.

11 Things IT Professionals Wish They Knew Earlier in Their Careers
Lisa Morgan, Freelance Writer,  4/6/2021
Time to Shift Your Job Search Out of Neutral
Jessica Davis, Senior Editor, Enterprise Apps,  3/31/2021
Does Identity Hinder Hybrid-Cloud and Multi-Cloud Adoption?
Joao-Pierre S. Ruth, Senior Writer,  4/1/2021
White Papers
Register for InformationWeek Newsletters
Current Issue
Successful Strategies for Digital Transformation
Download this report to learn about the latest technologies and best practices or ensuring a successful transition from outdated business transformation tactics.
Flash Poll