You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@linkis.apache.org by ca...@apache.org on 2022/09/02 03:20:34 UTC

[incubator-linkis] branch dev-1.3.1 updated: feat: scala code format alarm clear in linkis-storage (#3188)

This is an automated email from the ASF dual-hosted git repository.

casion pushed a commit to branch dev-1.3.1
in repository https://gitbox.apache.org/repos/asf/incubator-linkis.git


The following commit(s) were added to refs/heads/dev-1.3.1 by this push:
     new 4cdc4395c feat: scala code format alarm clear in linkis-storage (#3188)
4cdc4395c is described below

commit 4cdc4395ca87dfb473d99946131dda86459d60b7
Author: ruY <43...@users.noreply.github.com>
AuthorDate: Fri Sep 2 11:20:26 2022 +0800

    feat: scala code format alarm clear in linkis-storage (#3188)
---
 .../org/apache/linkis/storage/FSFactory.scala      |  3 +-
 .../org/apache/linkis/storage/domain/Dolphin.scala |  3 +-
 .../org/apache/linkis/storage/io/IOClient.scala    |  3 +-
 .../storage/io/IOMethodInterceptorCreator.scala    |  3 +-
 .../resultset/DefaultResultSetFactory.scala        | 12 +++--
 .../storage/resultset/StorageResultSetReader.scala |  6 ++-
 .../resultset/table/TableResultDeserializer.scala  | 11 ++--
 .../linkis/storage/script/ScriptFsWriter.scala     |  6 +--
 .../linkis/storage/script/VariableParser.scala     | 62 +++++++++++++---------
 .../storage/script/parser/CommonScriptParser.scala | 12 +++--
 .../script/reader/StorageScriptFsReader.scala      |  3 +-
 .../script/writer/StorageScriptFsWriter.scala      |  4 +-
 .../apache/linkis/storage/source/FileSplit.scala   |  6 +--
 .../linkis/storage/source/TextFileSource.scala     |  4 +-
 .../linkis/storage/utils/FileSystemUtils.scala     |  3 +-
 .../linkis/storage/utils/StorageHelper.scala       | 31 +++++++----
 16 files changed, 102 insertions(+), 70 deletions(-)

diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/FSFactory.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/FSFactory.scala
index 16e78309b..c61d33cfe 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/FSFactory.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/FSFactory.scala
@@ -31,8 +31,9 @@ object FSFactory extends Logging {
   )
 
   def getBuildFactory(fsName: String): BuildFactory = {
-    if (!buildClasses.contains(fsName))
+    if (!buildClasses.contains(fsName)) {
       throw new StorageFatalException(50000, s"Unsupported file system type(不支持的文件系统类型):$fsName")
+    }
     buildClasses(fsName)
   }
 
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/domain/Dolphin.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/domain/Dolphin.scala
index 33d898820..be9656eb1 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/domain/Dolphin.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/domain/Dolphin.scala
@@ -91,8 +91,9 @@ object Dolphin extends Logging {
   }
 
   def getType(content: String): String = {
-    if (content.length < MAGIC.length || content.substring(0, MAGIC.length) != MAGIC)
+    if (content.length < MAGIC.length || content.substring(0, MAGIC.length) != MAGIC) {
       throw new IOException(s"File header type must be dolphin,content:$content is not")
+    }
     content.substring(MAGIC.length, MAGIC.length + INT_LEN).toInt.toString
   }
 
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOClient.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOClient.scala
index 41ab3111a..0457b8a1f 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOClient.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOClient.scala
@@ -48,11 +48,12 @@ object IOClient {
   val FAILED = "FAILED"
 
   def getIOClient(): IOClient = {
-    if (ioClient == null)
+    if (ioClient == null) {
       throw new StorageErrorException(
         52004,
         "You must register IOClient before you can use proxy mode.(必须先注册IOClient,才能使用代理模式)"
       )
+    }
     ioClient
   }
 
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOMethodInterceptorCreator.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOMethodInterceptorCreator.scala
index a69be6bf6..ec298b517 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOMethodInterceptorCreator.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/io/IOMethodInterceptorCreator.scala
@@ -40,11 +40,12 @@ object IOMethodInterceptorCreator {
   }
 
   def getIOMethodInterceptor(fsName: String): MethodInterceptor = {
-    if (interceptorCreator == null)
+    if (interceptorCreator == null) {
       throw new StorageErrorException(
         52004,
         "You must register IOMethodInterceptorCreator before you can use proxy mode.(必须先注册IOMethodInterceptorCreator,才能使用代理模式)"
       )
+    }
     interceptorCreator.createIOMethodInterceptor(fsName)
   }
 
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/DefaultResultSetFactory.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/DefaultResultSetFactory.scala
index 60a47269c..f9fd3c9dc 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/DefaultResultSetFactory.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/DefaultResultSetFactory.scala
@@ -28,6 +28,7 @@ import org.apache.linkis.storage.utils.{StorageConfiguration, StorageUtils}
 import org.apache.commons.lang3.StringUtils
 
 import java.util
+import java.util.Locale
 
 class DefaultResultSetFactory extends ResultSetFactory with Logging {
 
@@ -35,14 +36,15 @@ class DefaultResultSetFactory extends ResultSetFactory with Logging {
     StorageUtils.loadClasses(
       StorageConfiguration.STORAGE_RESULT_SET_CLASSES.getValue,
       StorageConfiguration.STORAGE_RESULT_SET_PACKAGE.getValue,
-      t => t.newInstance().resultSetType().toLowerCase
+      t => t.newInstance().resultSetType().toLowerCase(Locale.getDefault)
     )
 
   val resultTypes = ResultSetFactory.resultSetType.keys.toArray
 
   override def getResultSetByType(resultSetType: String): ResultSet[_ <: MetaData, _ <: Record] = {
-    if (!resultClasses.contains(resultSetType))
+    if (!resultClasses.contains(resultSetType)) {
       throw new StorageErrorException(50000, s"Unsupported result type(不支持的结果类型):$resultSetType")
+    }
     resultClasses(resultSetType).newInstance()
   }
 
@@ -74,11 +76,12 @@ class DefaultResultSetFactory extends ResultSetFactory with Logging {
   override def getResultSetByPath(fsPath: FsPath, fs: Fs): ResultSet[_ <: MetaData, _ <: Record] = {
     val inputStream = fs.read(fsPath)
     val resultSetType = Dolphin.getType(inputStream)
-    if (StringUtils.isEmpty(resultSetType))
+    if (StringUtils.isEmpty(resultSetType)) {
       throw new StorageWarnException(
         51000,
         s"The file (${fsPath.getPath}) is empty(文件(${fsPath.getPath}) 为空)"
       )
+    }
     Utils.tryQuietly(inputStream.close())
     // Utils.tryQuietly(fs.close())
     getResultSetByType(resultSetType)
@@ -94,11 +97,12 @@ class DefaultResultSetFactory extends ResultSetFactory with Logging {
     fs.init(new util.HashMap[String, String]())
     val inputStream = fs.read(fsPath)
     val resultSetType = Dolphin.getType(inputStream)
-    if (StringUtils.isEmpty(resultSetType))
+    if (StringUtils.isEmpty(resultSetType)) {
       throw new StorageWarnException(
         51000,
         s"The file (${fsPath.getPath}) is empty(文件(${fsPath.getPath}) 为空)"
       )
+    }
     Utils.tryQuietly(inputStream.close())
     Utils.tryQuietly(fs.close())
     getResultSetByType(resultSetType)
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/StorageResultSetReader.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/StorageResultSetReader.scala
index 97075d6bd..3f49faf3e 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/StorageResultSetReader.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/StorageResultSetReader.scala
@@ -51,11 +51,12 @@ class StorageResultSetReader[K <: MetaData, V <: Record](
 
   def init(): Unit = {
     val resType = Dolphin.getType(inputStream)
-    if (resultSet.resultSetType != resType)
+    if (resultSet.resultSetType != resType) {
       throw new IOException(
         "File type does not match(文件类型不匹配): " + ResultSetFactory.resultSetType
           .getOrElse(resType, "TABLE")
       )
+    }
   }
 
   /**
@@ -96,10 +97,11 @@ class StorageResultSetReader[K <: MetaData, V <: Record](
   @scala.throws[IOException]
   override def getRecord: Record = {
     if (metaData == null) throw new IOException("Must read metadata first(必须先读取metadata)")
-    if (row == null)
+    if (row == null) {
       throw new IOException(
         "Can't get the value of the field, maybe the IO stream has been read or has been closed!(拿不到字段的值,也许IO流已读取完毕或已被关闭!)"
       )
+    }
     row
   }
 
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/table/TableResultDeserializer.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/table/TableResultDeserializer.scala
index db9ebbfb0..07612ef7e 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/table/TableResultDeserializer.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/resultset/table/TableResultDeserializer.scala
@@ -33,12 +33,13 @@ class TableResultDeserializer extends ResultDeserializer[TableMetaData, TableRec
     val colByteLen = Dolphin.getString(bytes, 0, Dolphin.INT_LEN).toInt
     val colString = Dolphin.getString(bytes, Dolphin.INT_LEN, colByteLen)
     val colArray =
-      if (colString.endsWith(Dolphin.COL_SPLIT))
+      if (colString.endsWith(Dolphin.COL_SPLIT)) {
         colString.substring(0, colString.length - 1).split(Dolphin.COL_SPLIT)
-      else colString.split(Dolphin.COL_SPLIT)
+      } else colString.split(Dolphin.COL_SPLIT)
     var index = Dolphin.INT_LEN + colByteLen
-    if (colArray.length % 3 != 0)
+    if (colArray.length % 3 != 0) {
       throw new StorageErrorException(52001, "Parsing metadata failed(解析元数据失败)")
+    }
     val columns = new ArrayBuffer[Column]()
     for (i <- 0 until (colArray.length, 3)) {
       var len = colArray(i).toInt
@@ -66,9 +67,9 @@ class TableResultDeserializer extends ResultDeserializer[TableMetaData, TableRec
     val colByteLen = Dolphin.getString(bytes, 0, Dolphin.INT_LEN).toInt
     val colString = Dolphin.getString(bytes, Dolphin.INT_LEN, colByteLen)
     val colArray =
-      if (colString.endsWith(Dolphin.COL_SPLIT))
+      if (colString.endsWith(Dolphin.COL_SPLIT)) {
         colString.substring(0, colString.length - 1).split(Dolphin.COL_SPLIT)
-      else colString.split(Dolphin.COL_SPLIT)
+      } else colString.split(Dolphin.COL_SPLIT)
     var index = Dolphin.INT_LEN + colByteLen
     val data = colArray.indices.map { i =>
       val len = colArray(i).toInt
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/ScriptFsWriter.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/ScriptFsWriter.scala
index eb2dde5ea..2de8687a2 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/ScriptFsWriter.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/ScriptFsWriter.scala
@@ -97,9 +97,9 @@ trait Compaction {
 class ScriptMetaData(var variables: Array[Variable]) extends MetaData {
   override def cloneMeta(): MetaData = new ScriptMetaData(variables)
 
-  def getMetaData = variables
+  def getMetaData: Array[Variable] = variables
 
-  def setMetaData(variables: Array[Variable]) = {
+  def setMetaData(variables: Array[Variable]): Unit = {
     this.variables = variables
   }
 
@@ -107,5 +107,5 @@ class ScriptMetaData(var variables: Array[Variable]) extends MetaData {
 
 class ScriptRecord(line: String) extends LineRecord(line)
 
-//definition  variable; specialConfiguration ;runConfiguration; startUpConfiguration;
+// definition  variable; specialConfiguration ;runConfiguration; startUpConfiguration;
 case class Variable(sortParent: String, sort: String, key: String, value: String)
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/VariableParser.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/VariableParser.scala
index c71204f54..ca89bd2f8 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/VariableParser.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/VariableParser.scala
@@ -30,25 +30,28 @@ object VariableParser {
   val SPECIAL: String = "special"
 
   def getVariables(params: util.Map[String, Object]): Array[Variable] = {
-    import scala.collection.JavaConversions._
+    import scala.collection.JavaConverters._
     val variables = new ArrayBuffer[Variable]
     params
       .getOrDefault(VARIABLE, new util.HashMap[String, Object])
       .asInstanceOf[util.Map[String, Object]]
+      .asScala
       .foreach(f => variables += Variable(VARIABLE, null, f._1, f._2.toString))
     params
       .getOrDefault(CONFIGURATION, new util.HashMap[String, Object])
       .asInstanceOf[util.Map[String, Object]]
+      .asScala
       .foreach { f =>
         f._2
           .asInstanceOf[util.Map[String, Object]]
+          .asScala
           .filter(s => !isContextIDINFO(s._1))
           .foreach(p =>
             p._2 match {
-              case e: util.Map[String, Object] => {
-                e.filter(s => !isContextIDINFO(s._1))
+              case e: util.Map[String, Object] =>
+                e.asScala
+                  .filter(s => !isContextIDINFO(s._1))
                   .foreach(s => variables += Variable(f._1, p._1, s._1, s._2.toString))
-              }
               case _ => variables += Variable(CONFIGURATION, f._1, p._1, p._2.toString)
             }
           )
@@ -62,41 +65,49 @@ object VariableParser {
   }
 
   def getMap(variables: Array[Variable]): util.Map[String, Object] = {
-    import scala.collection.JavaConversions._
+    import scala.collection.JavaConverters._
     val vars = new util.HashMap[String, String]
     val confs = new util.HashMap[String, Object]
-    variables.filter(_.sort == null).foreach(f => vars += f.key -> f.value)
+    variables.filter(_.sort == null).foreach(f => vars.asScala += f.key -> f.value)
     variables.filter(_.sort != null).foreach { f =>
       f.sort match {
         case STARTUP | RUNTIME | SPECIAL =>
-          if (confs.get(f.sort) == null)
-            confs += f.sort -> createMap(f)
-          else
-            confs.get(f.sort).asInstanceOf[util.HashMap[String, Object]] += f.key -> f.value
+          if (confs.get(f.sort) == null) {
+            confs.asScala += f.sort -> createMap(f)
+          } else {
+            confs.get(f.sort).asInstanceOf[util.HashMap[String, Object]].asScala += f.key -> f.value
+          }
         case _ =>
           if (confs.get(f.sortParent) == null) {
-            confs += f.sortParent -> new util.HashMap[String, Object]
+            confs.asScala += f.sortParent -> new util.HashMap[String, Object]
             confs
               .get(f.sortParent)
-              .asInstanceOf[util.HashMap[String, Object]] += f.sort -> createMap(f)
+              .asInstanceOf[util.HashMap[String, Object]]
+              .asScala += f.sort -> createMap(f)
           } else {
             val subMap = confs.get(f.sortParent).asInstanceOf[util.HashMap[String, Object]]
-            if (subMap.get(f.sort) == null) subMap += f.sort -> createMap(f)
-            else subMap.get(f.sort).asInstanceOf[util.HashMap[String, Object]] += f.key -> f.value
+            if (subMap.get(f.sort) == null) subMap.asScala += f.sort -> createMap(f)
+            else
+              subMap
+                .get(f.sort)
+                .asInstanceOf[util.HashMap[String, Object]]
+                .asScala += f.key -> f.value
           }
       }
     }
     val params = new util.HashMap[String, Object]
-    if (vars.size() > 0) params += VARIABLE -> vars
-    if (confs.size() > 0) params += CONFIGURATION -> confs
+    if (vars.size() > 0) params.asScala += VARIABLE -> vars
+    if (confs.size() > 0) params.asScala += CONFIGURATION -> confs
     params
   }
 
+  import scala.collection.JavaConverters._
+
   private def createMap(variable: Variable): util.Map[String, Object] = {
-    import scala.collection.JavaConversions._
+
     val map = new util.HashMap[String, Object]
-    map += variable.key -> variable.value
-  }
+    map.asScala += variable.key -> variable.value
+  }.asJava
 
   /* def main(args: Array[String]): Unit = {
     val a = Array(
@@ -106,11 +117,11 @@ object VariableParser {
       Variable("configuration", "runtime", "a", "b"),
       Variable("runtime", "env", "g2", "h3"),
       Variable("startup", "hello", "g2", "h3"))
-    //val a = Array[Variable]()
-    //println(new GsonBuilder().setDateFormat("yyyy-MM-dd'T'HH:mm:ssZ").serializeNulls.create.toJson(getMap(a)))
-    //val variables: Array[Variable] = getVariables(getMap(a))
-    //val variables: Array[Variable] = getVariables(getMap(a))
-    //print(variables)
+    // val a = Array[Variable]()
+    // println(new GsonBuilder().setDateFormat("yyyy-MM-dd'T'HH:mm:ssZ").serializeNulls.create.toJson(getMap(a)))
+    // val variables: Array[Variable] = getVariables(getMap(a))
+    // val variables: Array[Variable] = getVariables(getMap(a))
+    // print(variables)
    /* val writer = ScriptFsWriter.getScriptFsWriter(new FsPath("file:///tmp/hello.py"), "utf-8", new FileOutputStream("E:\\aaa.py"))
 
     writer.addMetaData(new ScriptMetaData(a))
@@ -123,7 +134,6 @@ object VariableParser {
     val map = getMap(reader.getMetaData.asInstanceOf[ScriptMetaData].getMetaData)
     println(new GsonBuilder().setDateFormat("yyyy-MM-dd'T'HH:mm:ssZ").serializeNulls.create.toJson(map))
     print(reader.getRecord.asInstanceOf[ScriptRecord].getLine)
-
-  }*/
+  } */
 
 }
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/parser/CommonScriptParser.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/parser/CommonScriptParser.scala
index 6f670c359..711b44315 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/parser/CommonScriptParser.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/parser/CommonScriptParser.scala
@@ -28,21 +28,23 @@ abstract class CommonScriptParser extends Parser {
     line match {
       case variableReg(key, value) =>
         Variable(VariableParser.VARIABLE, null, key.trim, value.trim)
-      case _ => {
+      case _ =>
         val split = line.split("=")
-        if (split.length != 2)
+        if (split.length != 2) {
           throw new StorageErrorException(65000, "Invalid custom parameter(不合法的自定义参数)")
+        }
         val value = split(1).trim
         val subSplit = split(0).split(" ")
-        if (subSplit.filter(_ != "").size != 4)
+        if (subSplit.filter(_ != "").size != 4) {
           throw new StorageErrorException(65000, "Invalid custom parameter(不合法的自定义参数)")
-        if (!subSplit.filter(_ != "")(0).equals(prefixConf))
+        }
+        if (!subSplit.filter(_ != "")(0).equals(prefixConf)) {
           throw new StorageErrorException(65000, "Invalid custom parameter(不合法的自定义参数)")
+        }
         val sortParent = subSplit.filter(_ != "")(1).trim
         val sort = subSplit.filter(_ != "")(2).trim
         val key = subSplit.filter(_ != "")(3).trim
         Variable(sortParent, sort, key, value)
-      }
     }
   }
 
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/reader/StorageScriptFsReader.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/reader/StorageScriptFsReader.scala
index 741589834..3f0613783 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/reader/StorageScriptFsReader.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/reader/StorageScriptFsReader.scala
@@ -101,13 +101,12 @@ class StorageScriptFsReader(val path: FsPath, val charset: String, val inputStre
     val regex = ("\\s*" + prefix + "\\s*(.+)\\s*" + "=" + "\\s*(.+)\\s*").r
     line match {
       case regex(_, _) => true
-      case _ => {
+      case _ =>
         val split: Array[String] = line.split("=")
         if (split.size != 2) return false
         if (split(0).split(" ").filter(_ != "").size != 4) return false
         if (!split(0).split(" ").filter(_ != "")(0).equals(prefixConf)) return false
         true
-      }
     }
   }
 
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/writer/StorageScriptFsWriter.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/writer/StorageScriptFsWriter.scala
index c415753c9..a318c5504 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/writer/StorageScriptFsWriter.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/script/writer/StorageScriptFsWriter.scala
@@ -53,8 +53,8 @@ class StorageScriptFsWriter(
       if (outputStream != null) {
         IOUtils.writeLines(metadataLine, "\n", outputStream, charset)
       } else {
-        import scala.collection.JavaConversions._
-        metadataLine.foreach(m => stringBuilder.append(s"$m\n"))
+        import scala.collection.JavaConverters._
+        metadataLine.asScala.foreach(m => stringBuilder.append(s"$m\n"))
       }
     }
   }
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/FileSplit.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/FileSplit.scala
index 731b7911a..0c7e3e7a1 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/FileSplit.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/FileSplit.scala
@@ -29,7 +29,7 @@ import org.apache.commons.math3.util.Pair
 import java.io.Closeable
 import java.util
 
-import scala.collection.JavaConversions._
+import scala.collection.JavaConverters._
 
 class FileSplit(
     var fsReader: FsReader[_ <: MetaData, _ <: Record],
@@ -63,7 +63,7 @@ class FileSplit(
   }
 
   def addParams(key: String, value: String): Unit = {
-    this.params += key -> value
+    this.params.asScala += key -> value
   }
 
   def `while`[M](m: MetaData => M, r: Record => Unit): M = {
@@ -133,7 +133,7 @@ class FileSplit(
       "comment" -> column.comment,
       "dataType" -> column.dataType.typeName
     )
-  }
+  }.asJava
 
   // 如果不分页,则一直读,如果分页,则 count需要小于count
   def ifContinueRead: Boolean = !pageTrigger || count <= end
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/TextFileSource.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/TextFileSource.scala
index 606c57f03..08ce8f51c 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/TextFileSource.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/source/TextFileSource.scala
@@ -24,7 +24,7 @@ import org.apache.commons.math3.util.Pair
 
 import java.util
 
-import scala.collection.JavaConversions._
+import scala.collection.JavaConverters._
 
 class TextFileSource(fileSplits: Array[FileSplit]) extends AbstractFileSource(fileSplits) {
 
@@ -39,7 +39,7 @@ class TextFileSource(fileSplits: Array[FileSplit]) extends AbstractFileSource(fi
     val snds: Array[util.ArrayList[Array[String]]] = collects.map(_.getSecond)
     snds.foreach { snd =>
       val str = new StringBuilder
-      snd.foreach {
+      snd.asScala.foreach {
         case Array("\n") => str.append("\n")
         case Array(y) => str.append(y).append("\n")
       }
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/FileSystemUtils.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/FileSystemUtils.scala
index c40c9a497..5252c12e0 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/FileSystemUtils.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/FileSystemUtils.scala
@@ -69,8 +69,9 @@ object FileSystemUtils extends Logging {
   ): Unit = {
     if (!fileSystem.exists(filePath)) {
       if (!fileSystem.exists(filePath.getParent)) {
-        if (!createParentWhenNotExists)
+        if (!createParentWhenNotExists) {
           throw new IOException("parent dir " + filePath.getParent.getPath + " dose not exists.")
+        }
         mkdirs(fileSystem, filePath.getParent, user)
       }
       fileSystem.createNewFile(filePath)
diff --git a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/StorageHelper.scala b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/StorageHelper.scala
index 3daf69430..7e5169fc5 100644
--- a/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/StorageHelper.scala
+++ b/linkis-commons/linkis-storage/src/main/scala/org/apache/linkis/storage/utils/StorageHelper.scala
@@ -17,14 +17,19 @@
 
 package org.apache.linkis.storage.utils
 
+import org.apache.linkis.common.utils.Logging
 import org.apache.linkis.storage.FSFactory
 import org.apache.linkis.storage.resultset.{ResultSetFactory, ResultSetReader}
 import org.apache.linkis.storage.resultset.table.{TableMetaData, TableRecord}
 
-object StorageHelper {
+/**
+ * 工具类,用于做storage jar包打出来做测试用 Tool class, which is used to print the storage jar package for testing
+ */
+
+object StorageHelper extends Logging {
 
   def main(args: Array[String]): Unit = {
-    if (args.length < 2) println("Usage method params eg:getTableResLines path")
+    if (args.length < 2) logger.info("Usage method params eg:getTableResLines path")
     val method = args(0)
     val params = args.slice(1, args.length)
     Thread.sleep(10000L)
@@ -33,7 +38,7 @@ object StorageHelper {
       case "getTableResLines" => getTableResLines(params)
       case "getTableRes" => getTableRes(params)
       case "createNewFile" => createNewFile(params)
-      case _ => println("There is no such method")
+      case _ => logger.info("There is no such method")
     }
   }
 
@@ -50,14 +55,14 @@ object StorageHelper {
     fs.init(null)
     val reader = ResultSetReader.getResultSetReader(resultSet, fs.read(resPath))
     val rmetaData = reader.getMetaData
-    rmetaData.asInstanceOf[TableMetaData].columns.foreach(println)
+    rmetaData.asInstanceOf[TableMetaData].columns.foreach(column => logger.info(column.toString))
     var num = 0
     Thread.sleep(10000L)
     while (reader.hasNext) {
       reader.getRecord
       num = num + 1
     }
-    println(num)
+    logger.info(num.toString)
     reader.close()
   }
 
@@ -71,8 +76,12 @@ object StorageHelper {
     fs.init(null)
     val reader = ResultSetReader.getResultSetReader(resultSet, fs.read(resPath))
     val rmetaData = reader.getMetaData
-    rmetaData.asInstanceOf[TableMetaData].columns.foreach(println)
-    rmetaData.asInstanceOf[TableMetaData].columns.map(_.columnName + ",").foreach(print)
+    rmetaData.asInstanceOf[TableMetaData].columns.foreach(column => logger.info(column.toString))
+    rmetaData
+      .asInstanceOf[TableMetaData]
+      .columns
+      .map(_.columnName + ",")
+      .foreach(column => logger.info(column))
     var num = 0
     while (reader.hasNext) {
       num = num + 1
@@ -80,10 +89,10 @@ object StorageHelper {
       if (num > len) {
         val record = reader.getRecord
         record.asInstanceOf[TableRecord].row.foreach { value =>
-          print(value.toString)
-          print(",")
+          logger.info(value.toString)
+          logger.info(",")
         }
-        println()
+        logger.info("\n")
       }
     }
   }
@@ -92,7 +101,7 @@ object StorageHelper {
     val resPath = StorageUtils.getFsPath(args(0))
     val proxyUser = StorageUtils.getJvmUser
     FileSystemUtils.createNewFile(resPath, proxyUser, true)
-    println("success")
+    logger.info("success")
   }
 
 }


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@linkis.apache.org
For additional commands, e-mail: commits-help@linkis.apache.org