Tag: Hadoop Tips

How to Install and Configure Hive with High Availability – Part 7

Hive is a Data Warehouse model in Hadoop Eco-System. It can perform as an ETL tool on top of Hadoop. Enabling High Availability (HA) on Hive is not similar as we do in Master Services

How to Set Up High Availability for Resource Manager – Part 6

YARN is the Processing Layer of Hadoop, which consists of the Master (Resource Manager) and Slave (Node Manager) services to process the data. Resource Manager (RM) is the critical component that is responsible for resource

How to Set Up High Availability for Namenode – Part 5

Hadoop has two core components which are HDFS and YARN. HDFS is for storing the Data, YARN is for processing the Data. HDFS is Hadoop Distributed File System, it has Namenode as Master Service and

How to Install CDH and Configure Service Placements on CentOS/RHEL 7 – Part 4

In an earlier article, we have explained the installation of Cloudera Manager, in this article, you will learn how to install and configure CDH (Cloudera Distribution Hadoop) in RHEL/CentOS 7. While installing the CDH parcel,

How to Install and Configure Cloudera Manager on CentOS/RHEL 7 – Part 3

In this article, we described the step by step process to install Cloudera Manager as per industrial practices. In Part 2, we already have gone through the Cloudera Pre-requisites, make sure all the servers are

Setting Up Hadoop Pre-requisites and Security Hardening – Part 2

Hadoop Cluster Building is a step by step process where the process starts from purchasing the required servers, mounting into the rack, cabling, etc. and placing in Datacentre. Then we need to install the OS,

Best Practices for Deploying Hadoop Server on CentOS/RHEL 7 – Part 1

In this series of articles, we are going to cover the entire Cloudera Hadoop Cluster Building building with Vendor and Industrial recommended best practices. Part 8: How to Install and Configure Sentry (Authorization Tool) Part

How to Install Hadoop Single Node Cluster (Pseudonode) on CentOS 7

Hadoop is an open-source framework that is widely used to deal with Bigdata. Most of the Bigdata/Data Analytics projects are being built up on top of the Hadoop Eco-System. It consists of two-layer, one is