You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kylin.apache.org by "wangrupeng (Jira)" <ji...@apache.org> on 2020/07/08 09:47:00 UTC

[jira] [Created] (KYLIN-4625) Debug the code of Kylin on Parquet without hadoop environment

wangrupeng created KYLIN-4625:
---------------------------------

             Summary: Debug the code of Kylin on Parquet without hadoop environment
                 Key: KYLIN-4625
                 URL: https://issues.apache.org/jira/browse/KYLIN-4625
             Project: Kylin
          Issue Type: Improvement
          Components: Spark Engine
            Reporter: wangrupeng
            Assignee: wangrupeng
         Attachments: image-2020-07-08-17-41-35-954.png, image-2020-07-08-17-42-09-603.png

Currently, Kylin on Parquet already supports debuging source code with local csv files, but it's a little bit complex. The steps are as follows:
* edit the properties of $KYLIN_SOURCE_DIR/examples/test_case_data/sandbox/kylin.properties to local
   ```log
   kylin.metadata.url=$LOCAL_META_DIR
   kylin.env.zookeeper-is-local=true
   kylin.env.hdfs-working-dir=file:///path/to/local/dir
   kylin.engine.spark-conf.spark.master=local
   kylin.engine.spark-conf.spark.eventLog.dir=/path/to/local/dir
   ```
* debug org.apache.kylin.rest.DebugTomcat with IDEA && add VM option "-Dspark.local=true" 
    !image-2020-07-08-17-41-35-954.png! 
* Load csv data source by pressing button "Data Source->Load CSV File as Table" on "Model" page, and set the schema for your table. Then press "submit" to save.
     !image-2020-07-08-17-42-09-603.png! 

Most time we debug just want to build and query cube easy. But current way is complex to load csv tables and create model and cube. So, I want to add a csv source  which using the model of kylin sample data directly when debug tomcat started.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)