From Zeppelin-0.7, Zeppelin started supporting Helium plugins/packages using Helium Framework. However, I am not able to view any of the plugin on Helium page (localhost:8080/#/helium). As per this JIRA, I placed sample Helium.json (available on s3) under /local-repo/helium-registry-cache. However, after that I got NPE while restarting Apache Zeppelin service.
I have tried Zeppelin 0.7 as well as Zeppelin 0.8.0 snaptshot versions. In particular, I want to use map Helium package - Helium-Map in Zeppelin note.
Can some one point me to any guide or documentation having detailed steps of using Helium package in Zeppelin? Any help would be greatly appreciated!
Zeppelin 0.7.x
Zeppelin 0.7.x doesn't support the online registry. In other words,
Zeppelin doesn't use helium.json. So you need to install each package by yourself.
clone the helium package what you want to install
modify the artifact value to the absolute path considering your local machine in helium-xxx.json
copy zeppelin-xxx.json into the $ZEPPELIN_HOME/helium directory (create if it doesn't exist yet)
restart Zeppelin and go to the localhost:8080/#/helium page, then install the package.
Zeppelin 0.8.0-SNAPSHOT
Zeppelin 0.8.0-SNAPSHOT supports the online registry. So you can install without any preparation.
But the NPE problem you've faced was fixed after https://github.com/apache/zeppelin/pull/2380.
So please git pull origin master and rebuild it :)
FYI, Now Zeppelin provides proxy functionality for helium. Refer https://github.com/apache/zeppelin/pull/2363
I work with Cloudera Manager CDH 5.7.1, which supports only Hive 1.1.0.
NiFi 1.0.0-BETA uses Hive 1.2.1.
When I try to use SelectHiveQL processor, I get the following error: Required field 'client_protocol' is unset!, which means that there's a version mismatch between Hive client and server.
Any suggestions to solve this problem?
I thought about building NiFi with hive-jdbc dependency version 1.1.0 instead of the default 1.2.1, but I hope there's a better solution.
Since NiFi is an Apache project, it builds with Apache JARs (such as Hive and Hadoop). However there are vendor-specific profiles and build properties you can use to build NiFi for a particular Hadoop distribution.
For example you could try the following to build a NiFi distro for CDH 5.7.1:
mvn clean install -DskipTests -Pcloudera -Dhadoop.version=2.6.0-cdh5.7.1 -Dhive.version=1.1.0-cdh5.7.1 -Dhbase.version=1.2.0-cdh5.7.1
The Hive processors use Hadoop libraries provided by the NiFi Hadoop Libraries NAR, and other NARs (like the Hadoop/HDFS processors) use this same libraries NAR, so the best approach is to build the whole thing. Otherwise you can try to replace just the Hadoop/Hive/HBase-related NARs and see if that works.
Because NiFi expects the newer version of Hive, it is necessary to remove the unsupported newer features (such as HiveStreaming and ORC support), support the older version of Thrift, and build against the Cloudera-specific libraries.
I have created a branch of the current NiFi-1.1.x release with the necessary changes to get the PutHiveQL and SelectHiveQL processors to work, which you could build as below:
git clone https://github.com/Chaffelson/nifi.git
git checkout nifi-1.1.x-cdhHiveBundle
mvn -T C2.0 clean install -Pcloudera -Dhive.version=1.1.0-cdh5.10.0 -Dhive.hadoop.version=2.6.0-cdh5.10.0 -Dhadoop.version=2.6.0-cdh5.10.0 -DskipTests
I have posted a more complete coverage of this on the Hortonworks Community forum: https://community.hortonworks.com/articles/93771/connecting-nifi-to-cdh-hive.html
I am currently using hadoop 1.0.3 version. I recently installed Apache Hive to run with it. I was running the select * query which gave me an NoSuchMethodError: org.apache.hadoop.mapred.JobConf.unset
I further found out its a compatibility issue with my current version of hadoop and requires me to upgrade to 1.2 or later.
I am fairly new to hadoop and would like to upgrade my current version to 1.2 or later. How do I go about doing the same.
I could not find any resources online to do so.
Thanks.
Just download hadoop 1.2.x from here and do necessary configuration changes in your new hadoop. Change HADOOP_HOME to point to your new hadoop folder.
NOTE: Change all the environmental variables (including .bashrc) to point to your new hadoop.
I am trying to install Apache ActiveMQ directly in Apache Karaf as per this link.
step 1: karaf#root()> feature:install jms
step 2: karaf#root()> feature:repo-add activemq
step 1& 2 sucessfully executed
step 3: karaf#root()> feature:install activemq-broker
step 3 shows the following error:
Error executing command: Unable to resolve root: missing requirement [root] osgi
.identity; osgi.identity=camel-jms; type=karaf.feature; version="[2.15.0,3.0.0)"
I am using apache karaf 4.0.0 m2 and apache activemq 5.9.0, please help me to resolve this problem
This doesn't work as per Karaf 4.0 the API provided is Servlet 3.1.
So you need to use a newer version of ActiveMQ a snapshot most likely.
AFAIK this issue has already been addressed at the Karaf user mailinglist.
The current and quick solution for this is to execute with Karaf 3.x. I got this working with Karaf 3.0.3. If you can downgrade Karaf, you may also have Karaf 3.x and install ActiveMQ features in it.
I'm attempting to build and run CouchDB following the instructions at https://github.com/jhs/build-couchdb. I am able to build successfully, but when I try to run CouchDB I first receive an error about a bad symbol lookup in a crypto library. Every subsequent time I run it I get the following error:
Apache CouchDB 1.6.0 (LogLevel=info) is starting.
{"init terminating in do_boot",{{badmatch,{error,{bad_return,{{couch_app,start,[normal,["/couchdb/build/etc/couchdb/default.ini","/couchdb/build/etc/couchdb/local.ini"]]},{'EXIT',{{badmatch,{error,{shutdown,{failed_to_start_child,couch_secondary_services,{shutdown,{failed_to_start_child,uuids,{undef,[{crypto,rand_bytes",[]},{couch_uuids,new_prefix,0,[{file,"/couchdb/dependencies/couchdb/src/couchdb/couch_uuids.erl"},{line,84}]},{couch_uuids,state,0,[{file,"/couchdb/dependencies/couchdb/src/couchdb/couch_uuids.erl"},{line,100}]},{couch_uuids,init,1,[{file,"/couchdb/dependencies/couchdb/src/couchdb/couch_uuids.erl"},{line,50}]},{gen_server,init_it,6,[{file,"gen_server.erl"},{line,304}]},{proc_lib,init_p_do_apply,3,[{file,"proc_lib.erl"},{line,239}]}]}}}}}}},[{couch_server_sup,start_server,1,[{file,"/couchdb/dependencies/couchdb/src/couchdb/couch_server_sup.erl"},{line,98}]},{application_master,start_it_old,4,[{file,"application_master.erl"},{line,269}]}]}}}}}},[{couch,start,0,[{file,"/couchdb/dependencies/couchdb/src/couchdb/couch.erl"},{line,18}]},{init,start_it,1,[]},{init,start_em,1,[]}]}}
I installed all of the dependencies listed on the github page as well as the dependencies listed on Apache's installation guide. Additionally, I have rebuilt with the option to not strip down erlang/otp
rake otp_keep="*"
I am running on CentOS 6.6
I ended up not using build-couchdb, but followed the instructions from the Apache wiki on installing couchdb. I had to set permissions as explained on the CouchDB guide, but it all worked out just fine.
This has been the best way I could figure out how to install CouchDB on Redhat/CentOS RHEL5/RHEL6.