You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "raj nelur (JIRA)" <ji...@apache.org> on 2017/04/17 04:20:41 UTC
[jira] [Created] (HIVE-16458) Inserting data into hive S3 external
table using INSERT INTO SELECT overwrites file
raj nelur created HIVE-16458:
--------------------------------
Summary: Inserting data into hive S3 external table using INSERT INTO SELECT overwrites file
Key: HIVE-16458
URL: https://issues.apache.org/jira/browse/HIVE-16458
Project: Hive
Issue Type: Bug
Components: Hive
Affects Versions: 1.2.1
Environment: Hadoop 2.7.3.2.5.3.0-37
Hive 1.2.1000.2.5.3.0-37
OS: RHEL 6.7
Reporter: raj nelur
I am trying to insert data into the external table using another table (tbl_test) which is not an external table.
insert into tbl_external select * from tbl_test;
But Hive is always generating a file with same file name (000000_0) when I run the above insert statement. As a result, the file is getting overwritten and the S3 location will always have a single file instead of multiple files when I run the insert statements multiple times.
How can I make Hive generate different file names? (eg 000000_0_copy_1, 000000_0_copy_2, etc)
We don't have this problem in our Dev environment, we are facing the problem only in Prod.
Dev Hive version is 1.2.1.2.3.6.0-3796
Prod Hive version is 1.2.1000.2.5.3.0-37
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)