After manipulating the imported records (for example, with Map Reduce or Hive) you may have a result data set which you can then back to the relational database.
Sqoop’s export process will read a set of delimited text files from HDFS in parallel, parse them into records, and insert them as new rows in a target database table, for consumption by external applications or users.
This class is used during the import process by Sqoop itself.
The Java source code for this class is also provided to you, for use in subsequent Map Reduce processing of the data.
If you are using Mac OS X, you should see few (if any) compatibility errors.
Sqoop is predominantly operated and tested on Linux. The input to the import process is a database table. The output of this import process is a set of files containing a copy of the imported table. For this reason, the output will be in multiple files.
Most aspects of the import, code generation, and export processes can be customized.
You can control the specific row range or columns imported.
You are also free to parse the delimiteds record data yourself, using any other tools you prefer.
When using Sqoop, the command line options that do not change from invocation to invocation can be put in an options file for convenience.
An options file is a text file where each line identifies an option in the order that it appears otherwise on the command line.
You can specify particular delimiters and escape characters for the file-based representation of the data, as well as the file format used.
You can also control the class or package names used in generated code.
This document is intended for: Before you can use Sqoop, a release of Hadoop must be installed and configured.