Run Apache Sqoop jobs by using .NET SDK for Apache Hadoop in HDInsight
Learn how to use the Azure HDInsight .NET SDK to run Apache Sqoop jobs in HDInsight to import and export between an HDInsight cluster and an Azure SQL Database or SQL Server database.
Prerequisites
Completion of Set up test environment from Use Apache Sqoop with Hadoop in HDInsight.
Familiarity with Sqoop. For more information, see Sqoop User Guide.
Use Sqoop on HDInsight clusters with the .NET SDK
The HDInsight .NET SDK provides .NET client libraries, so that it's easier to work with HDInsight clusters from .NET. In this section, you create a C# console application to export the hivesampletable
to the Azure SQL Database table that you created from the prerequisites.
Set up
Start Visual Studio and create a C# console application.
Navigate to Tools > NuGet Package Manager > Package Manager Console and run the following command:
Install-Package Microsoft.Azure.Management.HDInsight.Job
Sqoop export
From Hive to SQL Server. This example exports data from the Hive hivesampletable
table to the mobiledata
table in SQL Database.
Use the following code in the Program.cs file. Edit the code to set the values for
ExistingClusterName
, andExistingClusterPassword
.using Microsoft.Azure.Management.HDInsight.Job; using Microsoft.Azure.Management.HDInsight.Job.Models; using Hyak.Common; namespace SubmitHDInsightJobDotNet { class Program { private static HDInsightJobManagementClient _hdiJobManagementClient; private const string ExistingClusterName = "<Your HDInsight Cluster Name>"; private const string ExistingClusterPassword = "<Cluster User Password>"; private const string ExistingClusterUri = ExistingClusterName + ".azurehdinsight.cn"; private const string ExistingClusterUsername = "admin"; static void Main(string[] args) { System.Console.WriteLine("The application is running ..."); var clusterCredentials = new BasicAuthenticationCloudCredentials { Username = ExistingClusterUsername, Password = ExistingClusterPassword }; _hdiJobManagementClient = new HDInsightJobManagementClient(ExistingClusterUri, clusterCredentials); SubmitSqoopJob(); System.Console.WriteLine("Press ENTER to continue ..."); System.Console.ReadLine(); } private static void SubmitSqoopJob() { var sqlDatabaseServerName = ExistingClusterName + "dbserver"; var sqlDatabaseLogin = "sqluser"; var sqlDatabaseLoginPassword = ExistingClusterPassword; var sqlDatabaseDatabaseName = ExistingClusterName + "db"; // Connection string for using Azure SQL Database; Comment if using SQL Server var connectionString = "jdbc:sqlserver://" + sqlDatabaseServerName + ".database.chinacloudapi.cn;user=" + sqlDatabaseLogin + "@" + sqlDatabaseServerName + ";password=" + sqlDatabaseLoginPassword + ";database=" + sqlDatabaseDatabaseName; // Connection string for using SQL Server; Uncomment if using SQL Server // var connectionString = "jdbc:sqlserver://" + sqlDatabaseServerName + ";user=" + sqlDatabaseLogin + ";password=" + sqlDatabaseLoginPassword + ";database=" + sqlDatabaseDatabaseName; //sqoop start var tableName = "mobiledata"; var parameters = new SqoopJobSubmissionParameters { Command = "export --connect " + connectionString + " --table " + tableName + " --hcatalog-table hivesampletable" }; //sqoop end System.Console.WriteLine("Submitting the Sqoop job to the cluster..."); var response = _hdiJobManagementClient.JobManagement.SubmitSqoopJob(parameters); System.Console.WriteLine("Validating that the response is as expected..."); System.Console.WriteLine("Response status code is " + response.StatusCode); System.Console.WriteLine("Validating the response object..."); System.Console.WriteLine("JobId is " + response.JobSubmissionJsonResponse.Id); } } }
To run the program, select the F5 key.
Sqoop import
From SQL Server to Azure Storage. This example is dependent on the above export having been performed. This example imports data from the mobiledata
table in SQL Database to the wasb:///tutorials/usesqoop/importeddata
directory on the cluster's default Storage Account.
Replace the code above in the
//sqoop start //sqoop end
block with the following code:var tableName = "mobiledata"; var exportDir = "/tutorials/usesqoop/importeddata"; var parameters = new SqoopJobSubmissionParameters { Command = "import --connect " + connectionString + " --table " + tableName + " --target-dir " + exportDir + " --fields-terminated-by \\t --lines-terminated-by \\n -m 1" };
To run the program, select the F5 key.
Limitations
Linux-based HDInsight presents the following limitations:
Bulk export: The Sqoop connector that's used to export data to Microsoft SQL Server or Azure SQL Database doesn't currently support bulk inserts.
Batching: By using the
-batch
switch, Sqoop performs multiple inserts instead of batching the insert operations.
Next steps
Now you've learned how to use Sqoop. To learn more, see:
- Use Apache Oozie with HDInsight: Use Sqoop action in an Oozie workflow.
- Upload data to HDInsight: Find other methods for uploading data to HDInsight or Azure Blob storage.