You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Ken Dallmeyer (JIRA)" <ji...@apache.org> on 2014/03/07 23:57:42 UTC

[jira] [Created] (HIVE-6589) Automatically add partitions for external tables

Ken Dallmeyer created HIVE-6589:
-----------------------------------

             Summary: Automatically add partitions for external tables
                 Key: HIVE-6589
                 URL: https://issues.apache.org/jira/browse/HIVE-6589
             Project: Hive
          Issue Type: New Feature
    Affects Versions: 0.10.0
            Reporter: Ken Dallmeyer


I have a data stream being loaded into Hadoop via Flume. It loads into a date partition folder in HDFS.  The path looks like this:
{code}/flume/my_data/YYYY/MM/DD/HH
/flume/my_data/2014/03/02/01
/flume/my_data/2014/03/02/02
/flume/my_data/2014/03/02/03{code}

On top of it I create an EXTERNAL hive table to do querying.  As of now, I have to manually add partitions.  What I want is for EXTERNAL tables, Hive should "discover" those partitions.  Additionally I would like to specify a partition pattern so that when I query Hive will know to use the partition pattern to find the HDFS folder.

So something like this:
{code}CREATE EXTERNAL TABLE my_data (
  col1 STRING,
  col2 INT
)
PARTITIONED BY (
  dt STRING,
  hour STRING
)
LOCATION 
  '/flume/mydata'
TBLPROPERTIES (
  'hive.partition.spec' = 'dt=$Y-$M-$D, hour=$H',
  'hive.partition.spec.location' = '$Y/$M/$D/$H',
);
{code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)