Contact Us

Kockpit is here to help you

Business Form

Welcome to the Kockpit Community

Stay updated on the newest technologies, recent developments, and latest trends in Analytics and BI

DevOps Team

March 31

Get Started with the Built-In Image of Kockpit Hadoop and Spark Offering(StandAlone)

About:

Now you can enjoy the quick and hassle-free configuration of your workspaces with Kockpit's pre-configured images. Spark Standalone with Hadoop is a virtual machine image (VMI) created for Ubuntu and Red Hat (OS) that allows you to set up your machines within minutes.

For you, we have packed Spark and Hadoop in an image offering, enabling easy setup and configuration. With the help of an integrated form in KockpitShell.sh located in the /usr/local directory, you can connect to your Spark and Hadoop UIs without any need to develop scripts; you may create your standalone server just by filling out the form.

The distribution of Hadoop and Spark is based on Linux and is provided by Kockpit Analytics Pvt. Ltd. Hadoop and Spark Image is designed for Production Environments on Azure.

Follow the steps below to learn how to deploy Hadoop and Spark in Standalone Mode using Kockpit Image:

  1. Log in to your Azure VM using SSH

  2. After log in into your VM type the "ls" command in the /usr/local directory to get a list of files.

  3. Now as you can see there is a file named "KockpitShell.sh". Bash that file using the command "bash KockpitShell.sh"

    After Bashing the file, the following Screen will appear.

  4. In the first 4 fields fill the username, password, hostname of the machine and the Master IP of the VM. For example:

  5. In the last 2 fields, input the number of cores your machine has in Spark Cores and input the RAM your machine has in the Spark RAM field. For example my VM has 2 Cores and 4GB of RAM, so the input fields will be like this:

    Note: Make sure you add a small "g" to the Spark RAM field which stands for Giga Bytes.

  6. After filling up the above form, click on the submit button and wait for the script execution.

    The final output will be like this:

  7. Use the command "jps" to list all your active processes. Wait for 10 seconds and run jps again if the below processes do not show up. To successfully Setup a Hadoop and Spark Standalone cluster, the following processes should be running:

  8. Copy your VM IP and paste it in browser and port number in the IP to access the Web UI of Hadoop and Spark.

    Port For Hadoop is 9870 and for Spark it is 8080

    For Example:

    Your standalone server is ready to go!

Troubleshooting:

  • If you cannot view the same processes running in Step 7 or are unable to view the UI as represented in Step 8, you might have entered the wrong details in the form.

  • Enter the correct details in the Step 3 form.