Skip to content
This repository has been archived by the owner on Jul 9, 2021. It is now read-only.

SQOOP-324: Allow use of Avro Data Files with Hive import #75

Open
wants to merge 1 commit into
base: trunk
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion gradle.properties
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@

javaSourceCompatibilityVersion=1.8

avroVersion=1.8.1
avroVersion=1.8.2
parquetVersion=1.9.0
hadoopVersion=2.8.0
aspectjVersion=1.7.4
Expand Down
54 changes: 38 additions & 16 deletions src/java/org/apache/sqoop/hive/TableDefWriter.java
Original file line number Diff line number Diff line change
Expand Up @@ -21,6 +21,7 @@
import java.io.File;
import java.io.IOException;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collections;
import java.util.HashMap;
import java.util.List;
Expand Down Expand Up @@ -167,27 +168,36 @@ public String getCreateTableStmt() throws IOException {
}

boolean first = true;
String partitionKey = options.getHivePartitionKey();
for (String col : colNames) {
if (col.equals(partitionKey)) {
throw new IllegalArgumentException("Partition key " + col + " cannot "
+ "be a column to import.");
}
final String partitionKey = options.getHivePartitionKey();
if (Arrays.asList(colNames).contains(partitionKey)) {
throw new IllegalArgumentException(
"Partition key " + partitionKey + " cannot be a column to import.");
}
for (final String col : colNames) {

if (!first) {
sb.append(", ");
}

first = false;

String hiveColType;
if (options.getFileLayout() == SqoopOptions.FileLayout.TextFile) {
final String hiveColType;
switch (options.getFileLayout()) {
case TextFile:
case AvroDataFile:
Integer colType = columnTypes.get(col);
hiveColType = getHiveColumnTypeForTextTable(userMapping, col, colType);
} else if (options.getFileLayout() == SqoopOptions.FileLayout.ParquetFile) {
hiveColType = HiveTypes.toHiveType(columnNameToAvroFieldSchema.get(col), options);
} else {
throw new RuntimeException("File format is not supported for Hive tables.");
break;
case ParquetFile:
hiveColType =
HiveTypes.toHiveType(columnNameToAvroFieldSchema.get(col), options);
break;
case BinaryFile:
case SequenceFile:
default:
throw new RuntimeException(
"File format is not supported for Hive tables: "
+ options.getFileLayout());
}

sb.append('`').append(col).append("` ").append(hiveColType);
Expand All @@ -208,9 +218,14 @@ public String getCreateTableStmt() throws IOException {
.append(" STRING) ");
}

if (SqoopOptions.FileLayout.ParquetFile.equals(options.getFileLayout())) {
switch(options.getFileLayout()) {
case ParquetFile:
sb.append("STORED AS PARQUET");
} else {
break;
case AvroDataFile:
sb.append("STORED AS AVRO");
break;
case TextFile: {
sb.append("ROW FORMAT DELIMITED FIELDS TERMINATED BY '");
sb.append(getHiveOctalCharCode((int) options.getOutputFieldDelim()));
sb.append("' LINES TERMINATED BY '");
Expand All @@ -226,14 +241,21 @@ public String getCreateTableStmt() throws IOException {
sb.append("' STORED AS TEXTFILE");
}
}
break;
default:
throw new RuntimeException(
"File format is not supported for Hive tables: "
+ options.getFileLayout());
}

if (isHiveExternalTableSet) {
// add location
sb.append(" LOCATION '"+options.getHiveExternalTableDir()+"'");
}

LOG.debug("Create statement: " + sb.toString());
return sb.toString();
final String tableCreateStatement = sb.toString();
LOG.debug("Create statement: " + tableCreateStatement);
return tableCreateStatement;
}

private Map<String, Schema> getColumnNameToAvroTypeMapping() {
Expand Down
6 changes: 0 additions & 6 deletions src/java/org/apache/sqoop/tool/BaseSqoopTool.java
Original file line number Diff line number Diff line change
Expand Up @@ -1580,12 +1580,6 @@ protected void validateHiveOptions(SqoopOptions options)
+ " option." + HELP_STR);
}

if (options.doHiveImport()
&& options.getFileLayout() == SqoopOptions.FileLayout.AvroDataFile) {
throw new InvalidOptionsException("Hive import is not compatible with "
+ "importing into AVRO format.");
}

if (options.doHiveImport()
&& options.getFileLayout() == SqoopOptions.FileLayout.SequenceFile) {
throw new InvalidOptionsException("Hive import is not compatible with "
Expand Down
Loading