Hadoop and Hive Interview Cheat Sheet 1


SQL Based Datawarehouse app built on top of hadoop(select,join,groupby…..)

It is a platform used to develop SQL type scripts to do MapReduce operations.


Partition tables changes how HIVE structures the data storage
*Used for distributing load horizantally

ex: PARTITIONED BY (country STRING, state STRING);
A subset of a table’s data set where one column has the same value for all records in the subset.
In Hive, as in most databases that support partitioning, each partition is stored in a physically separate location—in Hive’s case, in a subdirectory of the root directory for the table.
Partitions have several advantages. The column value corresponding to a partition doesn’t have to be repeated
in every record in the partition, saving space, and queries with WHERE clauses that restrict the result set to specific values for the partition columns can perform more quickly, because they avoid scanning the directories of non matching partition values.

Partitioned by(country string,dept string)

A design that creates too many partitions is the large no of Hadoop files and directories that are created unnecessarily
Too many partitions may optimize some quieries but be detrimental
partitioned by(country string,state string);

hive> select * from partitionemployees where country=’CANADA’;

Dynamic ParttioningThis technique is convenient for
partitioning a query result into a potentially large number of partitions in a new table, without having to write a separate query for each partition column value.
SET hive.exec.dynamic.partition = true;
SET hive.exec.dynamic.partition.mode = nonstrict


IT is a technique for decomposing datasets into more manageable parts.
if we bucket table and use column as the bucketing column, the value of this column will be hashed by a user-defined number into buckets. Records with the same column will always be stored in the same bucket.

Decomposing datasets into more manageable parts while creating table clusteredby(uniqueid) into noof buckets No of buckets is fixed.
If 2 tables are bucketed by empid hive can create a logically correct sampling.it is also used for effiecient map side joins

External Partition Tables

  • It stores schema in a database and processed data into HDFS.
  • It is designed for OLAP.
  • It provides SQL type language for querying called HiveQL or HQL.
  • It is familiar, fast, scalable, and extensible.
  • Hive interfaces include WEB UI,
  • HIVE Command Line,
  • HD Insight(Windows Server)

External Table
Internal Table(Mangaed Table)
External Table
A table using a storage location and contents that are outside of Hive’s control.
It is convenient for sharing data with other tools,
when an external table is created, Hive does not create the external directory (or directories for partitioned tables), nor are the directory and data files deleted when an external table is

Data can be used outside of hive Hive cannot own control data.it delets only metadata Internal Tabe(or Manageable) Managed tables store their data inside the warehouse directory. Managed tables are less convenient for sharing with other tools. when you want full control over the table When u drp deletes data and schema in table Less convenient for sharing other tools.

Collection Data types


STRUCT {first STRING; last STRING}, then
the first name field can be referenced using name.first.
struct(‘John’, ‘Doe’)

arrays–similar data types–access by index
map-key value pairs–access by column[“federal”]
structs-different data types–access by dot
MAP A collection of key-value tuples, where the fields are accessed
using array notation (e.g., [‘key’]). For example, if a column
name is of type MAP with key?value pairs
‘first’?’John’ and ‘last’?’Doe’, then the last
name can be referenced using name[‘last’].
map(‘first’, ‘John’,’last’, ‘Doe’)

ARRAY Ordered sequences of the same type that are indexable using zero-based integers. For example, if a column name is of type ARRAY of strings with the value [‘John’, ‘Doe’], then the second element can be referenced using name[1].

The name is a simple string and for most employees, a float is large enough for the

Primitive data types

Complex struct

User Defined Aggregate Functions

User-defined functions that take multiple rows (or columns from multiple rows) and return a single “aggregation" of the data, such as a count of the rows, a sum or average of number values, etc.

User Defined Table generating Functions

User-defined functions that take a column from a single record and expand it into multiple rows.
Example includes converting a map into rows of key and value fields


Create new class extends udf Complie this class and add it to jar Put jar to hive classpath Register function


For speedy access to columns in your database Create index <INDEX_NAME> on table <TABLE_NAME> as name
Select * from <INDEX_NAME> where a = 1;


Hive uses ObjectInspector to analyze the internal structure of the row object and also the structure of the individual columns. ObjectInspector provides a uniform way to access complex objects that can be stored in
multiple formats in the memory, including:
• Instance of a Java class (Thrift or native Java)
• A standard Java object (we use java.util.List to represent Struct and Array, and use
java.util.Map to represent Map)
• A lazily-initialized object (For example, a Struct of string fields stored in a single Java string object with starting offset for each field) A complex object can be represented by a pair of ObjectInspector and Java Object. The ObjectInspector not only tells us the structure of the Object, but also gives us ways to access
the internal fields inside the Object.

create table

Copy Schema

create table if not exists customers.employeescopy like customers.employees;

–schema will be copied


The Serializer/Deserializer or SerDe for short is used to parse the bytes of a record into columns or fields, the deserialization process. It is also used to create those record bytes (i.e., serialization)

Built-in SerDes
Avro (Hive 0.9.1 and later)
ORC (Hive 0.11 and later)
Parquet (Hive 0.13 and later)
CSV (Hive 0.14 and later)

Serde is a library built on top of Hadoop API. Serializer, Deserializer instructs HIVE on how to process a record.
HIVE enables semi-structured or unstructured records to be processed also.
Serde will deseralize data from file to object so that it can be queried using SQL syntax.

Hive currently use these SerDe classes to serialize and deserialize data:
• MetadataTypedColumnsetSerDe: This SerDe is used to read/write delimited records like CSV, tab-separated control-A separated records (quote is not supported yet.)
• ThriftSerDe: This SerDe is used to read/write thrift serialized objects. The class file for the Thrift object must be loaded first.
• DynamicSerDe: This SerDe also read/write thrift serialized objects, but it understands thrift DDL so the schema of the object can be provided at runtime. Also it supports a lot of different protocols, including TBinaryProtocol, TJSONProtocol, TCTLSeparatedProtocol (which writes data in delimited records customserde

In most cases, users want to write a Deserializer instead of a SerDe, because users just want to read their own data format instead of writing to it.
•For example, the RegexDeserializer will deserialize the data using the configuration parameter ‘regex’, and possibly a list of column names
•If your SerDe supports DDL (basically, SerDe with parameterized columns and column types), you probably want to implement a Protocol based on DynamicSerDe, instead of writing a SerDe from scratch. The reason is that the framework passes DDL to SerDe through


  • An RPC(Remote Procedure Call) system integrated into Hive.
    Remote processes can send Hive statements to Hive through Thrift.
  • HIVE thrift allows access to HIVE over a single port.
    Thrift is a software framework for scalable cross language services development.
    Thrift allows clients using other languages to programmatically access HIVE remotely.
  • A Thrift service provides remote access from other processes.
  • Access using JDBC and ODBC are provided, too. They are implemented on top of the Thrift service

The service that maintains “metadata" information, such as table schemas.

Hive requires this service to be running.

support. By default, it uses a built-in Derby SQL server, which provides limited, single-process SQL for hive production environment,metastore service should run in an isolated jvm.
Hive processes can communicate with metastore services using thrift Hive metastore data is persisted in acid database such as db/mysql In hive-default.xml.template file we can set name to store in local or warehouse
<name> hive.metastore.warehouse.dir</name> <name>hive.metastore.local</name>

Hive metastore is a database that stores metadata about your Hive tables (eg. table name, column names and types, table location, storage handler being used, number of buckets in the table, sorting columns if any, partition columns if any, etc.). When you create a table, this metastore gets updated with the information related to the new table which gets queried when you issue queries on that table.


Avro is a serialization format developed to address some of the common problems associated with evolving other serialization formats.

Some of the benefits are:
rich data structures, fast binary format, support for remote procedure calls, and built-in schema evolution
JSON (JavaScript Object Notation) is a lightweight data serialization format used commonly in web-based applications.
(HDFS) A distributed, resilient file system for data storage (optimized for scanning large contiguous blocks of data on hard disks.) Distribution across a cluster provides horizontal scaling of data storage.

Blocks of HDFS files are replicated across the cluster (by default, three times) to prevent data loss when hard drives or whole servers fail.
Strict—these are used to protect large data in partitioned tables.Instead of extracting all data from partitioned table strict mode is safety measure which allows queries with where cause only on partitones

In the Hadoop context, a job is a self-contained workflow submitted to MapReduce.

It encompasses all the work required to perform a complete calculation, from reading input to generating output.

Job Tracker
The MapReduce JobTracker will decompose the job into one or more tasks for distribution and execution around the cluster.

The top-level controller of all jobs using Hadoop’s MapReduce.
The JobTracker accepts job submissions, determines what tasks to run and where to run them, monitors their execution, restarts failed tasks as needed, and provides a web console for monitoring job and task execution,
viewing logs, etc


In the MapReduce context, a task is the smallest unit of work performed on a single cluster node, as part of an overall job.
By default each task involves a separate JVM process.
Each map and reduce invocation will have its own task.

Input Format

The input format determines how input streams, usually from files, are split into records.

A SerDe handles parsing the record into columns. A custom input format can be specified when creating a table using the INPUTFORMAT clause.

Output Format

The output format determines how records are written to output streams, usually to files.
A SerDe handles serialization of each record into an appropriate byte stream.

A custom output format can be specified when creating a table using the OUTPUTFORMAT clause.

The output format for the default STORED AS TEXTFILE specification is implemented by the Java object named org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat


Partitioner controls the partitioning of the keys of the intermediate map-outputs. The key (or a subset of the key) is used to derive the partition, typically by a hash function. The total number of partitions is the same as the number of reduce tasks for the job.

hive stores schema and other Hive uses to store table schemas and other metadata for hive production environment,metastore service should run in an isolated jvm.
Hive processes can communicate with metastore services using thrift Hive metastore data is persisted in acid database such as db/mysql In hive-default.xml.template file we can set name to store in local or warehouse
<name> hive.metastore.warehouse.dir</name>