Unlock your Hadoop data with Hortonworks and Red Hat JBoss Data Virtualization

Welcome to this first episode of this series: “Unlock your [….] data with Red Hat JBoss Data Virtualization (JDV).”

This post will guide you through an example of connecting to a Hadoop source via the Hive2 driver, using Teiid Designer. In this example we will demonstrate connection to a local Hadoop source.  We’re using the Hortonworks 2.5 Sandbox running in Virtual Box for our source, but you can connect to another Hortonwork source if you wish using the same steps.

Hortonworks provides Hive JDBC and ODBC drivers that let you connect popular tools to query, analyze and visualize data stored within the Hortonworks Data Platform (HDP).

Note: we support HBase as well, stay tuned for an episode of Unlock your HBase data with Hortonworks and JDV.

Continue reading “Unlock your Hadoop data with Hortonworks and Red Hat JBoss Data Virtualization”

Share

Webinar: How to Stay Agile with Big Data: A Roadmap – 10 September

Agility is the key for benefiting from the use of Big Data for operational excellence and improved profitability. Ovum Research finds that organizations that take an iterative approach to refining analytic models, consolidating data sources, and transitioning to the cloud, tend to find more success with Big Data.

Attend this webinar to learn how to:

  • Consolidate your data sources
  • Build open, flexible Big Data ecosystems
  • Find success with Big Data

Continue reading “Webinar: How to Stay Agile with Big Data: A Roadmap – 10 September”

Share

DevNation 2014: Scott McClellan – Hadoop and Beyond

Abstract: Historically, the term “Hadoop” has been considered synonymous with its core technologies: MapReduce and the Hadoop Distributed File System (HDFS). But today the definition of Hadoop is rapidly evolving.

The Hadoop community is generalizing the application runtime model beyond MapReduce. On the storage front, we’re seeing the emergence of many alternative Hadoop-compatible file systems. Red Hat has built an interface layer for its Red Hat Storage Server product. This complete implementation of the Hadoop file system interface lets Hadoop-related projects run transparently, directly on a Red Hat Storage Server cluster.

Continue reading “DevNation 2014: Scott McClellan – Hadoop and Beyond”

Share