How to install cloudera on top of apache hadoop 2.7.1 - apache

I am currently working on apache hadoop2.7.1, cluster includes 1 name node and 3 data nodes.
Is it possible to install cloudera manager on existing apache hadoop 2.7.1 cluster. If yes, could you please suggest me how it can be done.
Thanks in advance.

No, that is not possible. It goes the other way around. You install Cloudera Manager, and then you deploy Hadoop components from its web console.

Related

Is it posible to install Hortonworks Cluster on Docker on one single Linux machine

We want to build test Hadoop cluster on one Linux machine based on docker container/s
Dose hortonworks ( cloudera ) support this ? ,
For example HDP version - 2.6.5
For example we need the following services
HDFS ( include at least 3 data-nodes , )
YARN
MAPreduce2
HIVE
Zookeeper
Ambari metrics
KAFKA
SPARK2
and all these services should be on one Linux machine
#Jessica it is not possible to do a 3 node (hdfs) instance with a single node machine. Other than that, a single node cluster can run all of the services you have listed and is suitable for learning, training, and demos or proof of concepts.

Configure Redis Cluster in Ubuntu Server 14.04

I've installed redis-server using apt-get install redis-server and everything went fine.
Right now I'm trying to configure it in a Cluster mode. The problem is that in the tutorial supplied here http://redis.io/topics/cluster-tutorial they use a script called redis-trib.rb which I can't find it in my system.
Can you please tell me how can I configure my Redis to run in Cluster mode without that script ?
I would like to have a setup with two masters, each on a different machine.
Thank you very much.
Had same problem with reredis-trib.rb
This tutorial explains how to create Redis Cluster using only Redis commands: Configuring and Running Redis Cluster on Linux
You need Redis 3.0.0 beta to run Cluster. You'll not find it in a Linux distribution, since they all have copy of the stable server (fortunately!). Redis 3.0.0 will go out as a stable release the next week. You can find the source code of the stable release here: http://redis.io/download.
There is now a tutorial for Ubuntu at https://www.digitalocean.com/community/tutorials/how-to-configure-a-redis-cluster-on-ubuntu-14-04 which includes installation of a PPA to supply 3.0.x. This tutorial is only for two nodes and does not reference redis-trib.rb ...

Ambari setup on top of Hadoop

I have the following already setup (in order)
A Hadoop 2.5.x - 3 nodes cluster.
Apache Ambari Version 1.7.0
So my question is, Is there anyway I can integrate/configure Amabri to monitor my already setup Hadoop cluster ?
Any suggestions/views are much appreciate.
Currently it seems like that's not possible, there is open JIRA ticket on this issue:
https://issues.apache.org/jira/browse/AMBARI-2852

Can Presto connect to other Hadoop distributions and run queries

I see Presto has plugin only to CDH4. Can I connect to other distributions such as HortonWorks from this and what does it take to do it.
Without a specific plugin, I am running into "path host null" errors when executing queries from Presto. Appreciate your help.
The Presto Hive connector supports multiple versions of Hadoop:
hive-hadoop1: Apache Hadoop 1.x
hive-hadoop2: Apache Hadoop 2.x
hive-cdh4: Cloudera CDH 4
hive-cdh5: Cloudera CDH 5
See the Hive Connector documentation for more details.
Where is the code for the CDH connector in GitHub?
briefly looking at the code in GitHub, i dont see anything specific to CDH , other than the name, in presto / presto-hive-cdh4 /src / main /java - am i looking at the wrong thing?

Accessing Hadoop clusters from eclipse

I just followed the Hadoop(0.20.2) installation tutorial and did the set up. I can run map reduce program on the cluster through eclipse. Now my problem is how can I connect to Hadoop clusters from my local system. Local system is windows 7 and I have installed eclipse plugin for Hadoop. I was trying to connect to Hadoop from my local system which is windows(My local system and Hadoop system are in same subnet). I got connection timed out error while connecting to Hadoop server.
In configuration files of Hadoop I have given actual IP addresses.
Not sure which step I have missed out?
I recently read, that the eclipse plugin won't work at all. But you can simply connect to your Cluster with the configuration keys:
mapred.job.tracker
fs.default.name
EDIT: here is a working version Apache Jira: Eclipse Plugin does not work with Eclipse Ganymede (3.4)