EducationSoftwareStrategy.com
StrategyCommunity

Knowledge Base

Product

Community

Knowledge Base

TopicsBrowse ArticlesDeveloper Zone

Product

Download SoftwareProduct DocumentationSecurity Hub

Education

Tutorial VideosSolution GalleryEducation courses

Community

GuidelinesGrandmastersEvents
x_social-icon_white.svglinkedin_social-icon_white.svg
Strategy logoCommunity

© Strategy Inc. All Rights Reserved.

LegalTerms of UsePrivacy Policy
  1. Home
  2. Topics

Hadoop (HDFS + MapReduce) Homepage


Stefan Zepeda

Salesforce Solutions Architect • Strategy


The MicroStrategy platform empowers organizations to build applications that leverage big data and Hadoop distributions.

The Strategy platform empowers organizations to build applications that leverage big data and Hadoop distributions. All of the major Hadoop distributions are certified to work with Strategy, and once connected, data stored in Hadoop becomes just like any other data source.
Users can connect using Hive, Pig, or proprietary SQL-on-Hadoop connectors like Cloudera, Impala, or IBM BigInsights.

ka02R000000kbFAQAY_0EM440000002JgL.jpeg

The Strategy Hadoop Gateway can natively tap into HDFS, generating schema on-read and making Hadoop suitable for ad-hoc querying. It also enables parallel loading of data from HDFS, resulting in high performance data loading.
Strategy's native connectivity saves users from the tedious process of ETL from HDFS to Hive and enhances performance in comparison to retrieving data from Hive when using ODBC.

ka02R000000kbFAQAY_0EM440000002JgQ.png

When to use ODBC connection to Hadoop?
Traditional SQL based access for reporting and dashboarding uses Project Schema and Developer to build models on top of Hadoop. It is also possible to build reports, documents and dashboards via live connect or in-memory cubes.  ODBC is the preferred method if your design requirements include:
•    Leverage Hadoop layer security at runtime
•    Live connection
•    A project schema is required
When to use the Hadoop Gateway?
High-performance, parallelized native access to Hadoop which uses Web's Data Import functionality to publish in-memory cubes for access and modeling.
It can also be used to build reports, documents and dashboards on top of in-memory cubes.  This is the preferred method if  your design requirements include:
•    Bulk data transfer into memory
•    Data wrangling
•    Browse and preview Hadoop files via import GUI
 


Comment

0 comments

Details

Knowledge Article

Published:

April 18, 2017

Last Updated:

September 6, 2017