I am happy to announce that Splunk Hadoop Connect 1.1 is now available. This version of Hadoop Connect rounds out Splunk’s integration with the Hadoop distributions by becoming certified on MapR. Cloudera, Hortonworks, and Apache Hadoop distributions also have the ability to benefit from the power of Splunk.
Splunk Hadoop Connect provides bi-directional integration to easily and reliably move data between Splunk and Hadoop. It provides Hadoop users the ability to gain real-time analysis, visualization and role based access control for a stream of machine-generated data. It delivers three core capacities: Export data from Splunk to Hadoop, Explore Hadoop directories and Import data from Hadoop to Splunk.
The most significant new feature added to version 1.1 is the ability to select whether you want to map to a remote HDFS cluster or to a mounted file system. The option to map to a mounted file system enables Splunk to integrate with MapR Hadoop distribution. MapR allows users to mount Hadoop via NFS using a feature called direct access NFS.
The three HDFS options available for users are:
1) Remote HDFS
** For example, ‘ namenode.hadoop.example.com:8020 ‘
** Supported distributions include Apache Hadoop, Cloudera, and Hortonworks
2) Remote HDFS with Kerberos Security
** For example, hdfs/namenode.hadoop.example.com@example.com
3) Locally mounted HDFS
** For example, ‘ /mnt/hadoop ‘
** Supported distributions include MapR and any other distribution that uses NFS mounted files.
For additional information about Splunk Hadoop Connect read these blogs:
http://blogs.splunk.com/2012/12/20/connecting-splunk-and-hadoop/
http://blogs.splunk.com/2012/11/28/hadoop-and-splunk-use-cases/
http://blogs.splunk.com/2012/08/27/do-you-hadoop-how-splunk-can-help/