HDFS is the distributed file system used in Hadoop and helps to achieve the purpose of storing very larger files on a commodity Hardware. While working on Hadoop and BigData in general it is very important to understand the basic concepts of they underlying file system, i.e. HDFS in case of Hadoop. When you are appearing in BigData Interviews , … Continue Reading ››
BigData and Data analytics Jobs are the most sought after jobs of current time. It is important to understand the basics before you appear for interview. In this post, I am covering few of the basic MapReduce interview questions for Hadoop MapReduce.
- What is MapReduce ?
- What is combiner and when you … Continue Reading ››
Resilient Distributed Datasets -RDDs in SparkApcahe Spark has already taken over Hadoop (MapReduce) because of plenty of benefits it provides in terms of faster execution in iterative processing algorithms such as Machine learning. In this post, we will try to understand what makes Spark RDDs so useful in batch analytics .
Why RDD ?When it comes to iterative distributed computing, i.e. processing data … Continue Reading ››
What is HCatalog ?Apache HCatalog is a Storage Management Layer for Hadoop that helps to users of different data processing tools in Hadoop ecosystem like Hive, Pig and MapReduce easily read and write data from the cluster.HCatalog enables with relational view of data from RCFile format, Parquet, ORC files, Sequence files stored on HDFS. It also exposes REST API exposed … Continue Reading ››
Since Hadoop jobs are often long running, its difficult for newbies to manage the processes in Unix unless they know some useful Unix commands to do so, so that they can increase their efficiency. In this post, I will explain some of the commands that are very useful while executing some long running jobs .We will see how to execute a … Continue Reading ››