Skip to main content

All in one HDFS Cluster for your pocket

Update 1 (Nov 21, 2011):
- added 3rd interface as host-only-adapter (hadoop1)
- enabled trusted device eth2

About one year ago, I created a small XEN-environment for my engineering pourposes. When I was traveling for hours it was very helpful to track some issues or test new features. The problem was that I had to carry 2 notebooks with me. That was the reason I switched to VirtualBox [1] which runs on OSX, Linux and Windows as well. I could play with my servers and when I did, they configured to death and I reimported them into a clean setup. I think that will also be a good start for new people who have to find into the hadoop ecosystem to see the power without the harm of configuration in a multi-node environment.
The appliance is created with VirtualBox, because it runs on OSX and Windows very easily. The idea behind it is to check new settings in a small environment rather easily; the appliance is designed for research, not for development and really not for production. The appliance has 4 nodes, one master and 3 slaves. The setup is not perfect, but it matched the environment I created it for. We have no seperate secondary namenode, for example. I set up hdfs, hive with mysql-metastore, hBase in distributed mode with zookeeper and stargate.

Before we can play with our own LAB we have to consider that we need some specials before. Please read the site [2] I created for.



  1. Pretty interesting. This is a good way to create an Hadoop test environment and actually our team is going to use it. I currently use VMWare Player to do something similar on one box, to get a full cluster up for testing purposes. I am the lead developer of, an Hadoop management software tool and so its important to have a test environment that is portable to I can test things quickly.

  2. I was testing with vmware-player, but I missed some features VirtualBox provides. The first is the transparency, I can use the app with OSX, Windows (7) and Linux as well.

    For consulting is really cool - you can demonstrate some changes in seconds live.

  3. Thanks a lot for this contribution! Based on this I could prepare my test environment in just some minutes. I also tested with vmware-player and finally I switched to VirtualBox too, which runs now on Windows7 and OpenSuse12.1.

    What do you think about a git-hub repository to collect useful admin and/or developer scripts which can than be deployed to a clean preinstalled DEMO- TEST- or TRAINING-cluster which can be based on your work?

  4. @Mirko: Sounds like a good idea, especially for ant builds I think.

  5. I created the repository here ...

  6. Hello Sir,
    I am student & need your help on below error
    Hadoop Error while running in multinode cluster

    root@ubuntu:/opt/hadoop-1.0.0# bin/hadoop jar hadoop-examples-1.0.0.jar pi 10 1$

    Number of Maps = 10
    Samples per Map = 10
    12/02/03 09:01:47 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:48 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:49 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:50 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:51 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:52 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:53 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:54 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$
    12/02/03 09:01:55 INFO ipc.Client: Retrying connect to server: ubuntu/192.168.1$

    Please help me if anyone can have solution on this error
    hadoop 1.0
    Ubuntu 11.10
    jdk 1.7


Post a Comment

Popular posts from this blog

Deal with corrupted messages in Apache Kafka

Under some strange circumstances it can happen that a message in a Kafka topic is corrupted. This happens often by using 3rd party frameworks together with Kafka. Additionally, Kafka < 0.9 has no lock at at the consumer read level, but has a lock on Log.write(). This can cause a rare race condition, as described in KAKFA-2477 [1]. Probably a log entry looks like: ERROR Error processing message, stopping consumer: ($) kafka.message.InvalidMessageException: Message is corrupt (stored crc = xxxxxxxxxx, computed crc = yyyyyyyyyy Kafka-Tools Kafka stores the offset of every consumer in Zookeeper. To read out the offsets, Kafka provides handy tools [2]. But also can be used, at least to display the consumer and the stored offsets. First we need to find the consumer for a topic (> Kafka 0.9): bin/ --zookeeper management01:2181 --describe --group test Prior to Kafka 0.9 the only possibility to get this inform

Hive query shows ERROR "too many counters"

A hive job face the odd " Too many counters:"  like Ended Job = job_xxxxxx with exception 'org.apache.hadoop.mapreduce.counters.LimitExceededException(Too many counters: 201 max=200)' FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MapRedTask Intercepting System.exit(1) These happens when operators are used in queries ( Hive Operators ). Hive creates 4 counters per operator, max upto 1000, plus a few additional counters like file read/write, partitions and tables. Hence the number of counter required is going to be dependent upon the query.  To avoid such exception, configure " mapreduce.job.counters.max " in mapreduce-site.xml to a value above 1000. Hive will fail when he is hitting the 1k counts, but other MR jobs not. A number around 1120 should be a good choice. Using " EXPLAIN EXTENDED " and " grep -ri operators | wc -l " print out the used numbers of operators. Use this value to tweak the MR s

GPT & GenAI for Startup Storytelling

OpenAI and Bard   are the most used GenAI tools today; the first one has a massive Microsoft investment, and the other one is an experiment from Google. But did you know that you can also use them to optimize and hack your startup?  For startups, creating pitch scripts, sales emails, and elevator pitches with generative AI (GenAI) can help you not only save time but also validate your marketing and wording. Curious? Here are a few prompt hacks for startups to create,improve, and validate buyer personas, your startup's mission/vision statements, and unique selling proposition (USP) definitions. First Step: Introduce yourself and your startup Introduce yourself, your startup, your website, your idea, your position, and in a few words what you are doing to the chatbot: Prompt : I'm NAME and our startup NAME, with website URL, is doing WHATEVER. With PRODUCT NAME, we aim to change or disrupt INDUSTRY. Bard is able to pull information from your website. I'm not sure if ChatGPT