Sqoop

APACHE SQOOP TUTORIAL

Tool To Transfer Data Between Apache Hadoop & Structured Datastores

Sqoop Complete Tutorial Part-1

Sqoop (SQL-to-Hadoop) is a Hadoop ecosystem component and an ETL tool that offers the capability to extract data from various structured data stores such as relational databases with the help of map-reduce. This command-line interpreter works efficiently to transfer a huge volume of the data to the Hadoop file system (HDFS, HBase, and Hive). Similarly,

Continue reading

Sqoop Complete Tutorial Part-2

This is the continuation part of “Sqoop Complete Tutorial”. If you want to read Part-1, please click here. Part-2, please click here. Part-3, please click here. Part-4, please click here. Part-5, please click here. Part-6, please click here. Part-7, please click here. We are about to discuss the following in the current and in the

Continue reading

Sqoop Complete Tutorial Part-3

This is the continuation part of “Sqoop Complete Tutorial”. If you want to read –Part-1, please click here.Part-2, please click here.Part-3, please click here.Part-4, please click here.Part-5, please click here.Part-6, please click here.Part-7, please click here. 5) Importing all the tables but excluding few tables from MySQL to HDFS  sqoop import-all-tables–connect jdbc:mysql://localhost/retail_db–username root–password cloudera–exclude-tables “order_items” 

Continue reading

Sqoop Complete Tutorial Part-4

This is the continuation part of “Sqoop Complete Tutorial”. If you want to read – Part-1, please click here.Part-2, please click here.Part-3, please click here.Part-4, please click here.Part-5, please click here.Part-6, please click here.Part-7, please click here. 11) Importing data from MySQL to HDFS based on SQL query  sqoop import-all-tables–connect jdbc:mysql://localhost/retail_db–username root–password cloudera–exclude-tables “order_items”  Append

Continue reading

Sqoop Complete Tutorial Part-5

This is the continuation part of “Sqoop Complete Tutorial”. If you want to read – Part-1, please click here. Part-2, please click here. Part-3, please click here. Part-4, please click here. Part-5, please click here. Part-6, please click here. Part-7, please click here. 14) Read the contents of the database without importing it.  sqoop list-tables

Continue reading

Sqoop Complete Tutorial Part-6

This is the continuation part of “Sqoop Complete Tutorial”. If you want to read – Part-1, please click here. Part-2, please click here. Part-3, please click here. Part-4, please click here. Part-5, please click here. Part-6, please click here. Part-7, please click here. 18) Importing a table from MySQL to Hive’s default database.   Importing a

Continue reading

Sqoop Complete Tutorial Part-7

This is the continuation part of “Sqoop Complete Tutorial”. If you want to read – Part-1, please click here. Part-2, please click here. Part-3, please click here. Part-4, please click here. Part-5, please click here. Part-6, please click here. Part-7, please click here. 20) Importing all tables from MySQL to Hive   Importing a table from

Continue reading

SQOOP Complete Tutorial Part-9

Previous articles covered how to move data from a relational database to HDFS and Hive. We’ll now look at how to get the data out of HDFS and Hive. Before exporting, the target table must exist in the target database. As previously stated, everything in HDFS is a file; there are no tables. When these

Continue reading

SQOOP Complete Tutorial Part-10

In this session, we’ll work with staging tables while exporting data from HDFS to MySQL. Staging tables are the tables that hold the data temporarily. If any exceptions are raised by the map reduce jobs, the data may only be loaded partly in the target table. Instead of transferring data to the target table, staging

Continue reading

SQOOP Complete Tutorial Part-11

This is the continuation part of “Sqoop Complete Tutorial”. The following will be covered in this article. Sqoop Export – Update the records in the destination table. Sqoop Export – Append the new and modify the existing the values in the destination table. When exporting data from Hive/HDFS to RDBMS, we frequently need to either

Continue reading

Sqoop Complete Tutorial Part-8

This is the continuation part of “Sqoop Complete Tutorial”. 22) Import the table’s structure from MySQL to Hive without data Create-hive-table is the option that generates the table structure based on the table available in the relational data source. sqoop create-hive-table –connect jdbc:mysql://localhost/empdept –username root –password cloudera –table employees –hive-table emp; Note: The above command

Continue reading

Sqoop Complete Tutorial Part-12

So far, we’ve looked at a variety of instances in which the same arguments are supplied several times. The arguments are mostly what you need to know when dealing with Sqoop. If the correct inputs are supplied, the required task will be completed without problems. This post is part of the “Sqoop Complete Tutorial” series,

Continue reading

Sqoop Complete Tutorial Part-13

This is continuation part to “SQOOP Complete Tutorial” and in this part we will see how to create and save the jobs in Sqoop. Instead of rewriting commands that need to be run many times, Sqoop allows you to save them as jobs. When it comes to incremental import, this is a common occurrence. A

Continue reading


Follow My Blog

Get new content delivered directly to your inbox.