Home:ALL Converter>Is it recommended to utilise the Hadoop Ecosystems parallel processing to manage and maintain slowly changing data warehouse dimensions?

Is it recommended to utilise the Hadoop Ecosystems parallel processing to manage and maintain slowly changing data warehouse dimensions?

Ask Time:2019-03-31T20:18:11         Author:norm

Json Formatter

We have transactional hourly data and daily bulk loaded data from operational systems which we engineer into data warehouse dimensions (and facts) in a MS SQL Server environment. If recommended, what fast and efficient techniques (Flume, Sqoop, Kafka, HDFS, HBase, Hive, KUDU, Spark, Impla) could we apply to achieve this in Hadoop? The assumption is that these dimensions would be consumed hourly by the MS SQL Server data warehouse.

Author:norm,eproduced under the CC 4.0 BY-SA copyright license with a link to the original source and this disclaimer.
Link to original article:https://stackoverflow.com/questions/55440809/is-it-recommended-to-utilise-the-hadoop-ecosystems-parallel-processing-to-manage
yy