You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@carbondata.apache.org by xm_zzc <44...@qq.com> on 2017/04/14 17:23:20 UTC
java.io.FileNotFoundException:
file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Hi all:
Please help. I directly ran a CarbonData demo program on Eclipse, which
copy from
carbondata-examples-spark2/src/main/scala/org/apache/carbondata/examples/CarbonSessionExample.scala,
but the error occurred, as follows:
*Exception in thread "main" java.io.FileNotFoundException:
file:/data/carbon_data/default/carbon_table/Metadata/schema.write *
at java.io.FileOutputStream.open0(Native Method)
at java.io.FileOutputStream.open(FileOutputStream.java:270)
at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
at java.io.FileOutputStream.<init>(FileOutputStream.java:101)
at
org.apache.carbondata.core.datastore.impl.FileFactory.getDataOutputStream(FileFactory.java:188)
at
org.apache.carbondata.core.fileoperations.AtomicFileOperationsImpl.openForWrite(AtomicFileOperationsImpl.java:61)
at
org.apache.carbondata.core.writer.ThriftWriter.open(ThriftWriter.java:97)
at
org.apache.spark.sql.hive.CarbonMetastore.createSchemaThriftFile(CarbonMetastore.scala:412)
at
org.apache.spark.sql.hive.CarbonMetastore.createTableFromThrift(CarbonMetastore.scala:380)
at
org.apache.spark.sql.execution.command.CreateTable.run(carbonTableSchema.scala:166)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at
org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
at
org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
at
org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87)
at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
at cn.xm.zzc.carbontest.FirstCarbonData$.main(FirstCarbonData.scala:51)
at cn.xm.zzc.carbontest.FirstCarbonData.main(FirstCarbonData.scala)
but I found a file named 'schema.write' in the path
'/data/carbon_data/default/carbon_table/Metadata/', the file size is 0.
My program :
/val warehouseLocation = Constants.SPARK_WAREHOUSE
val storeLocation = Constants.CARBON_FILES
//CarbonProperties.getInstance()
// .addProperty("carbon.storelocation", storeLocation)
CarbonProperties.getInstance()
.addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT,
"yyyy/MM/dd")
import org.apache.spark.sql.CarbonSession._
val spark = SparkSession
.builder()
.appName("FirstCarbonData")
.master("local")
.config("spark.sql.warehouse.dir", warehouseLocation)
//.config("javax.jdo.option.ConnectionURL",
// s"jdbc:derby:;databaseName=${Constants.METASTORE_DB};create=true")
//.enableHiveSupport()
.getOrCreateCarbonSession(storeLocation, Constants.METASTORE_DB)
spark.sql("DROP TABLE IF EXISTS carbon_table")
// Create table
spark.sql(
s"""
| CREATE TABLE carbon_table(
| shortField short,
| intField int,
| bigintField long,
| doubleField double,
| stringField string,
| timestampField timestamp,
| decimalField decimal(18,2),
| dateField date,
| charField char(5),
| floatField float,
| complexData array<string>
| )
| STORED BY 'carbondata'
| TBLPROPERTIES('DICTIONARY_INCLUDE'='dateField, charField')
""".stripMargin)
val path =
"/home/myubuntu/Works/workspace_latest/incubator-carbondata/examples/spark2/src/main/resources/data.csv"
// scalastyle:off
spark.sql(
s"""
| LOAD DATA LOCAL INPATH '$path'
| INTO TABLE carbon_table
|
options('FILEHEADER'='shortField,intField,bigintField,doubleField,stringField,timestampField,decimalField,dateField,charField,floatField,complexData','COMPLEX_DELIMITER_LEVEL_1'='#')
""".stripMargin)
// scalastyle:on
spark.sql("""
SELECT *
FROM carbon_table
where stringfield = 'spark' and decimalField > 40
""").show
spark.sql("""
SELECT *
FROM carbon_table where length(stringField) = 5
""").show
spark.sql("""
SELECT *
FROM carbon_table where date_format(dateField, "yyyy-MM-dd") =
"2015-07-23"
""").show
spark.sql("""
select count(stringField) from carbon_table
""".stripMargin).show
spark.sql("""
SELECT sum(intField), stringField
FROM carbon_table
GROUP BY stringField
""").show
spark.sql(
"""
|select t1.*, t2.*
|from carbon_table t1, carbon_table t2
|where t1.stringField = t2.stringField
""".stripMargin).show
spark.sql(
"""
|with t1 as (
|select * from carbon_table
|union all
|select * from carbon_table
|)
|select t1.*, t2.*
|from t1, carbon_table t2
|where t1.stringField = t2.stringField
""".stripMargin).show
spark.sql("""
SELECT *
FROM carbon_table
where stringfield = 'spark' and floatField > 2.8
""").show
// Drop table
// spark.sql("DROP TABLE IF EXISTS carbon_table")
spark.stop()/
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/java-io-FileNotFoundException-file-data-carbon-data-default-carbon-table-Metadata-schema-write-tp11044.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: java.io.FileNotFoundException:
file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by xm_zzc <44...@qq.com>.
Hi:
https://github.com/apache/incubator-carbondata/pull/813 this pr has
resolved this problem. Now it works fine.
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/java-io-FileNotFoundException-file-data-carbon-data-default-carbon-table-Metadata-schema-write-tp11044p11278.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: java.io.FileNotFoundException: file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by Bhavya Aggarwal <bh...@knoldus.com>.
Hi ,
Since you are running the example it is an issue but normally when you run
carbon data you only provide the carbon.store location and it accepts the
normal HDFS path e.g for running thrift server with carbon I generally use
the below command
/usr/local/spark-2.1/bin/spark-submit --conf
spark.sql.hive.thriftServer.singleSession=true --class
org.apache.carbondata.spark.thriftserver.CarbonThriftServer
/usr/local/spark-2.1/carbonlib/carbondata_2.11-1.1.0-incubating-SNAPSHOT-shade-hadoop2.2.0.jar
*hdfs://localhost:54311/user/hive/warehouse/carbon21.store*
where Highlighted part is the HDFS path.
I am currently looking into the problem with your dictionary file and will
let you know.
Regards
Bhavya
On Tue, Apr 18, 2017 at 1:06 PM, xm_zzc <44...@qq.com> wrote:
> Hi:
> After removing 'file://', the error 'FileNotFoundException' disappeared
> when loading data into table. But when I execute select query, there is an
> error: 'java.io.IOException: Dictionary file does not exist:
> /data/carbon_data/default/carbon_table/Metadata/3ca9a876-31e0-44cd-8497-
> fd190dbcd352.dictmeta'.
> I can't find this file in path:
> '/data/carbon_data/default/carbon_table/Metadata/'
>
> $ ll /data/carbon_data/default/carbon_table/Metadata/
> Total 52
> drwxrwxr-x 2 myubuntu myubuntu 4096 4月 18 15:01 ./
> drwxrwxr-x 4 myubuntu myubuntu 4096 4月 18 15:01 ../
> -rw-rw-r-- 1 myubuntu myubuntu 15 4月 18 15:01
> 803c1c9a-56ba-43a3-847c-ae4a83dfcbe0_36.sortindex
> -rw-rw-r-- 1 myubuntu myubuntu 36 4月 18 15:01
> 803c1c9a-56ba-43a3-847c-ae4a83dfcbe0.dict
> -rw-rw-r-- 1 myubuntu myubuntu 11 4月 18 15:01
> 803c1c9a-56ba-43a3-847c-ae4a83dfcbe0.dictmeta
> -rw-rw-r-- 1 myubuntu myubuntu 27 4月 18 15:01
> ba631ef5-b01c-4556-a286-550bbc98be80_52.sortindex
> -rw-rw-r-- 1 myubuntu myubuntu 52 4月 18 15:01
> ba631ef5-b01c-4556-a286-550bbc98be80.dict
> -rw-rw-r-- 1 myubuntu myubuntu 11 4月 18 15:01
> ba631ef5-b01c-4556-a286-550bbc98be80.dictmeta
> -rw-rw-r-- 1 myubuntu myubuntu 13 4月 18 15:01
> e42fa72f-43dc-4a58-9a5f-0e23d1653665_32.sortindex
> -rw-rw-r-- 1 myubuntu myubuntu 32 4月 18 15:01
> e42fa72f-43dc-4a58-9a5f-0e23d1653665.dict
> -rw-rw-r-- 1 myubuntu myubuntu 11 4月 18 15:01
> e42fa72f-43dc-4a58-9a5f-0e23d1653665.dictmeta
> -rw-rw-r-- 1 myubuntu myubuntu 1397 4月 18 15:01 schema
> -rw-rw-r-- 1 myubuntu myubuntu 276 4月 18 15:01 tablestatus
>
> What't the problem with this error?
>
> BTW, Why i can not use the 'file://' prefix, how to configure above paths
> on
> HDFS when I use HDFS?
>
> Thanks.
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/java-io-
> FileNotFoundException-file-data-carbon-data-default-
> carbon-table-Metadata-schema-write-tp11044p11234.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>
Re: java.io.FileNotFoundException:
file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by xm_zzc <44...@qq.com>.
Hi:
After removing 'file://', the error 'FileNotFoundException' disappeared
when loading data into table. But when I execute select query, there is an
error: 'java.io.IOException: Dictionary file does not exist:
/data/carbon_data/default/carbon_table/Metadata/3ca9a876-31e0-44cd-8497-fd190dbcd352.dictmeta'.
I can't find this file in path:
'/data/carbon_data/default/carbon_table/Metadata/'
$ ll /data/carbon_data/default/carbon_table/Metadata/
Total 52
drwxrwxr-x 2 myubuntu myubuntu 4096 4月 18 15:01 ./
drwxrwxr-x 4 myubuntu myubuntu 4096 4月 18 15:01 ../
-rw-rw-r-- 1 myubuntu myubuntu 15 4月 18 15:01
803c1c9a-56ba-43a3-847c-ae4a83dfcbe0_36.sortindex
-rw-rw-r-- 1 myubuntu myubuntu 36 4月 18 15:01
803c1c9a-56ba-43a3-847c-ae4a83dfcbe0.dict
-rw-rw-r-- 1 myubuntu myubuntu 11 4月 18 15:01
803c1c9a-56ba-43a3-847c-ae4a83dfcbe0.dictmeta
-rw-rw-r-- 1 myubuntu myubuntu 27 4月 18 15:01
ba631ef5-b01c-4556-a286-550bbc98be80_52.sortindex
-rw-rw-r-- 1 myubuntu myubuntu 52 4月 18 15:01
ba631ef5-b01c-4556-a286-550bbc98be80.dict
-rw-rw-r-- 1 myubuntu myubuntu 11 4月 18 15:01
ba631ef5-b01c-4556-a286-550bbc98be80.dictmeta
-rw-rw-r-- 1 myubuntu myubuntu 13 4月 18 15:01
e42fa72f-43dc-4a58-9a5f-0e23d1653665_32.sortindex
-rw-rw-r-- 1 myubuntu myubuntu 32 4月 18 15:01
e42fa72f-43dc-4a58-9a5f-0e23d1653665.dict
-rw-rw-r-- 1 myubuntu myubuntu 11 4月 18 15:01
e42fa72f-43dc-4a58-9a5f-0e23d1653665.dictmeta
-rw-rw-r-- 1 myubuntu myubuntu 1397 4月 18 15:01 schema
-rw-rw-r-- 1 myubuntu myubuntu 276 4月 18 15:01 tablestatus
What't the problem with this error?
BTW, Why i can not use the 'file://' prefix, how to configure above paths on
HDFS when I use HDFS?
Thanks.
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/java-io-FileNotFoundException-file-data-carbon-data-default-carbon-table-Metadata-schema-write-tp11044p11234.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: java.io.FileNotFoundException: file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by Bhavya Aggarwal <bh...@knoldus.com>.
Please use the below properties and don't append file:// and try again.
val METASTORE_DB: String = "/data/carbon_meta"
val SPARK_WAREHOUSE: String = "/data/spark-warehouse-carbon"
val CARBON_FILES: String = "/data/carbon_data"
Regards
Bhavya
On Mon, Apr 17, 2017 at 3:24 PM, xm_zzc <44...@qq.com> wrote:
> Hi:
> val METASTORE_DB: String = "/data/carbon_meta"
> val SPARK_WAREHOUSE: String = "file:///data/spark-warehouse-carbon"
> val CARBON_FILES: String = "file:///data/carbon_data"
>
> The permissions on above paths:
> drwxrwxrwx 3 myubuntu myubuntu 4096 4月 17 17:50 /data/carbon_data/
> drwxrwxr-x 3 myubuntu myubuntu 4096 4月 17 17:50
> /data/carbon_data/default/
> drwxrwxr-x 3 myubuntu myubuntu 4096 4月 17 17:50
> /data/carbon_data/default/carbon_table/
> drwxrwxr-x 2 myubuntu myubuntu 4096 4月 17 17:50
> /data/carbon_data/default/carbon_table/Metadata/
> -rw-rw-r-- 1 myubuntu myubuntu 0 4月 17 17:50
> /data/carbon_data/default/carbon_table/Metadata/schema.write
> drwxrwxrwx 3 myubuntu myubuntu 4096 4月 17 17:50 carbon_meta/
> drwxrwxrwx 2 myubuntu myubuntu 4096 4月 17 17:49
> spark-warehouse-carbon/
>
>
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/java-io-
> FileNotFoundException-file-data-carbon-data-default-
> carbon-table-Metadata-schema-write-tp11044p11178.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>
Re: java.io.FileNotFoundException:
file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by xm_zzc <44...@qq.com>.
Hi:
val METASTORE_DB: String = "/data/carbon_meta"
val SPARK_WAREHOUSE: String = "file:///data/spark-warehouse-carbon"
val CARBON_FILES: String = "file:///data/carbon_data"
The permissions on above paths:
drwxrwxrwx 3 myubuntu myubuntu 4096 4月 17 17:50 /data/carbon_data/
drwxrwxr-x 3 myubuntu myubuntu 4096 4月 17 17:50
/data/carbon_data/default/
drwxrwxr-x 3 myubuntu myubuntu 4096 4月 17 17:50
/data/carbon_data/default/carbon_table/
drwxrwxr-x 2 myubuntu myubuntu 4096 4月 17 17:50
/data/carbon_data/default/carbon_table/Metadata/
-rw-rw-r-- 1 myubuntu myubuntu 0 4月 17 17:50
/data/carbon_data/default/carbon_table/Metadata/schema.write
drwxrwxrwx 3 myubuntu myubuntu 4096 4月 17 17:50 carbon_meta/
drwxrwxrwx 2 myubuntu myubuntu 4096 4月 17 17:49
spark-warehouse-carbon/
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/java-io-FileNotFoundException-file-data-carbon-data-default-carbon-table-Metadata-schema-write-tp11044p11178.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: java.io.FileNotFoundException: file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by Bhavya Aggarwal <bh...@knoldus.com>.
Hi,
Can you please provide me the details of path for the following
Constants.CARBON_FILES
Constants.METASTORE_DB
Also please let me know the permissions on /data/carbon_data directory,
does the user you are using to run the example has permissions on directory
/data or /data/carbondata
Regards
Bhavya
On Mon, Apr 17, 2017 at 1:34 PM, xm_zzc <44...@qq.com> wrote:
> Hi Liang Chen:
> Thanks for your reply.
> I have added the '777' right to paths of 'Constants.CARBON_FILES' and
> 'Constants.METASTORE_DB'. but the error still exists. I have not idea how
> to
> resolve.
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/java-io-
> FileNotFoundException-file-data-carbon-data-default-
> carbon-table-Metadata-schema-write-tp11044p11154.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>
Re: java.io.FileNotFoundException:
file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by xm_zzc <44...@qq.com>.
Hi Liang Chen:
Thanks for your reply.
I have added the '777' right to paths of 'Constants.CARBON_FILES' and
'Constants.METASTORE_DB'. but the error still exists. I have not idea how to
resolve.
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/java-io-FileNotFoundException-file-data-carbon-data-default-carbon-table-Metadata-schema-write-tp11044p11154.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: java.io.FileNotFoundException:
file:/data/carbon_data/default/carbon_table/Metadata/schema.write
Posted by Liang Chen <ch...@gmail.com>.
Hi
Please check if you have the right for the directory: Constants.METASTORE_DB
you can use "chmod" to add right.
Regards
Liang
xm_zzc wrote
> Hi all:
> Please help. I directly ran a CarbonData demo program on Eclipse, which
> copy from
> carbondata-examples-spark2/src/main/scala/org/apache/carbondata/examples/CarbonSessionExample.scala,
> but the error occurred, as follows:
*
> Exception in thread "main" java.io.FileNotFoundException:
> file:/data/carbon_data/default/carbon_table/Metadata/schema.write
*
> at java.io.FileOutputStream.open0(Native Method)
> at java.io.FileOutputStream.open(FileOutputStream.java:270)
> at java.io.FileOutputStream.
> <init>
> (FileOutputStream.java:213)
> at java.io.FileOutputStream.
> <init>
> (FileOutputStream.java:101)
> at
> org.apache.carbondata.core.datastore.impl.FileFactory.getDataOutputStream(FileFactory.java:188)
> at
> org.apache.carbondata.core.fileoperations.AtomicFileOperationsImpl.openForWrite(AtomicFileOperationsImpl.java:61)
> at
> org.apache.carbondata.core.writer.ThriftWriter.open(ThriftWriter.java:97)
> at
> org.apache.spark.sql.hive.CarbonMetastore.createSchemaThriftFile(CarbonMetastore.scala:412)
> at
> org.apache.spark.sql.hive.CarbonMetastore.createTableFromThrift(CarbonMetastore.scala:380)
> at
> org.apache.spark.sql.execution.command.CreateTable.run(carbonTableSchema.scala:166)
> at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
> at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
> at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
> at
> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87)
> at org.apache.spark.sql.Dataset.
> <init>
> (Dataset.scala:185)
> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
> at cn.xm.zzc.carbontest.FirstCarbonData$.main(FirstCarbonData.scala:51)
> at cn.xm.zzc.carbontest.FirstCarbonData.main(FirstCarbonData.scala)
>
> but I found a file named 'schema.write' in the path
> '/data/carbon_data/default/carbon_table/Metadata/', the file size is 0.
>
> My program :
>
>
/
> val warehouseLocation = Constants.SPARK_WAREHOUSE
> val storeLocation = Constants.CARBON_FILES
>
> //CarbonProperties.getInstance()
> // .addProperty("carbon.storelocation", storeLocation)
> CarbonProperties.getInstance()
> .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT,
> "yyyy/MM/dd")
>
> import org.apache.spark.sql.CarbonSession._
> val spark = SparkSession
> .builder()
> .appName("FirstCarbonData")
> .master("local")
> .config("spark.sql.warehouse.dir", warehouseLocation)
> //.config("javax.jdo.option.ConnectionURL",
> //
> s"jdbc:derby:;databaseName=${Constants.METASTORE_DB};create=true")
> //.enableHiveSupport()
> .getOrCreateCarbonSession(storeLocation, Constants.METASTORE_DB)
>
> spark.sql("DROP TABLE IF EXISTS carbon_table")
>
> // Create table
> spark.sql(
> s"""
> | CREATE TABLE carbon_table(
> | shortField short,
> | intField int,
> | bigintField long,
> | doubleField double,
> | stringField string,
> | timestampField timestamp,
> | decimalField decimal(18,2),
> | dateField date,
> | charField char(5),
> | floatField float,
> | complexData array
> <string>
> | )
> | STORED BY 'carbondata'
> | TBLPROPERTIES('DICTIONARY_INCLUDE'='dateField, charField')
> """.stripMargin)
>
> val path =
> "/home/myubuntu/Works/workspace_latest/incubator-carbondata/examples/spark2/src/main/resources/data.csv"
>
> // scalastyle:off
> spark.sql(
> s"""
> | LOAD DATA LOCAL INPATH '$path'
> | INTO TABLE carbon_table
> |
> options('FILEHEADER'='shortField,intField,bigintField,doubleField,stringField,timestampField,decimalField,dateField,charField,floatField,complexData','COMPLEX_DELIMITER_LEVEL_1'='#')
> """.stripMargin)
> // scalastyle:on
>
> spark.sql("""
> SELECT *
> FROM carbon_table
> where stringfield = 'spark' and decimalField > 40
> """).show
>
> spark.sql("""
> SELECT *
> FROM carbon_table where length(stringField) = 5
> """).show
>
> spark.sql("""
> SELECT *
> FROM carbon_table where date_format(dateField, "yyyy-MM-dd")
> = "2015-07-23"
> """).show
>
> spark.sql("""
> select count(stringField) from carbon_table
> """.stripMargin).show
>
> spark.sql("""
> SELECT sum(intField), stringField
> FROM carbon_table
> GROUP BY stringField
> """).show
>
> spark.sql(
> """
> |select t1.*, t2.*
> |from carbon_table t1, carbon_table t2
> |where t1.stringField = t2.stringField
> """.stripMargin).show
>
> spark.sql(
> """
> |with t1 as (
> |select * from carbon_table
> |union all
> |select * from carbon_table
> |)
> |select t1.*, t2.*
> |from t1, carbon_table t2
> |where t1.stringField = t2.stringField
> """.stripMargin).show
>
> spark.sql("""
> SELECT *
> FROM carbon_table
> where stringfield = 'spark' and floatField > 2.8
> """).show
>
> // Drop table
> // spark.sql("DROP TABLE IF EXISTS carbon_table")
>
> spark.stop()
/
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/java-io-FileNotFoundException-file-data-carbon-data-default-carbon-table-Metadata-schema-write-tp11044p11087.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.